Information Theory

Chapter 1

1.1 Entropy of a source

A source is an ordered pair , where is a finite set, known as a source alphabet, and is a probability distribution over . Denote the probability of by or .

A function satisfy properties 1-3 if and only if it has the form

where and if .

Let be a probability distribution. Then is the b-ary entropy of the distribution .

If is the distribution of a source, is the entropy of the source.

1.2 Properties of entropy

Chapter 2

2.1 variable length encoding

2.2 Huffman Encoding

One of the optimal r-ary encoding scheme with minimum average length.

2.3 Noiseless Coding Theorem

Chapter 3 Noise Coding

3.1 Discrete Memoryless Channel and Coditional Entropy

Binary symmetric channel:
Binary erasure channel:

3.2 Mutual Information and Channel Capacity

The definition is the same for random vectors and .

capacity:
lossless:
deterministic:
noiseless:
useless:
binary symmetric channel: , p is crossover probability

3.3 Noisy Coding Theorem

A decision scheme is a partial function from the set of output strings to the set of codewords.
"partial" means the function is only defined for part of output strings.
decision error/decoding error: when is not the codeword that had been sent.

Assume the codeword is , and the output string received is .
The probability f a decision error:

Idea observers depends on the input distributions.
For uniform input, it is a maximum likelihood decision scheme.

Chapter 4 General remarks on codes

4.1 Error detection and correction

other definitions:
binary symmetric channel, crossover probability, channel matrix
input distribution


output distribution

joint probability

backward channel probability

decision scheme, decision error/decoding error, idea observer, average probability of error, maximum likelihood decision, the Noisy Coding Theorem.

4.2 Minimum Distance Decoding

minimum distance decoding: choose a codeword that is closest to the received word (maximum likelihood decoding).

A -code is a code with length , size , and minimum distance .

t-error-correcting: at most errors can be corrected with MDD
exactly t-error-correcting

4.3 Families of codes

Types of codes

For linear code, .

Families of codes

  1. repetition codes
    -ary repetition code of length is
  2. Hamming codes
    is a -ary -code, single-error-correcting
  3. Galay codes

Chapter 5: Linear Codes

5.1 Linear codes and dual

The codewords in are the linear combinations of thr rows of .
Standard form: , where is the identity matrix of size .
It is systematic on its first coordinate positions.

The quotient space of modulo is


a coset of :

standard array

0

The first row is the codewords in . The -th row is formed by choosing a word of smallest weight that is not yet in the array, and adding it to each word of the first row, to form the coset . is the coset leaders.

syndrome decoding: if is received, compute its syndrome, find the coset leader with the same syndrome, and decode as .

5.3 Maximum Distance Separable Codes

If a code is MDS, then so is its dual code.

The support of a vector is the set of all coordinate positions where is nonzero.