41
Shannon’s Noisy-Channel Coding Theorem Lucas Slot Sebastian Zur February 13, 2015 Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 1 / 29

Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

  • Upload
    others

  • View
    6

  • Download
    1

Embed Size (px)

Citation preview

Page 1: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Shannon’s Noisy-Channel Coding Theorem

Lucas Slot Sebastian Zur

February 13, 2015

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 1 / 29

Page 2: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Outline

1 Definitions and TerminologyDiscrete Memoryless ChannelsTerminologyJointly Typical Sets

2 Noisy-Channel Coding TheoremStatementPart onePart twoPart three

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 2 / 29

Page 3: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Outline

1 Definitions and TerminologyDiscrete Memoryless ChannelsTerminologyJointly Typical Sets

2 Noisy-Channel Coding TheoremStatementPart onePart twoPart three

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 3 / 29

Page 4: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Discrete Memoryless Channels

Definition

A discrete memoryless channel consist of two random variables X and Yover finite discrete alphabets X and Y that satisfy

P(X = x ,Y = y) = P(X = x)P(Y = y |X = x) ∀x , y ∈ X × Y

where X is the input and Y is the output of the channel.

Example

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 4 / 29

Page 5: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Discrete Memoryless Channels

Definition

A discrete memoryless channel consist of two random variables X and Yover finite discrete alphabets X and Y that satisfy

P(X = x ,Y = y) = P(X = x)P(Y = y |X = x) ∀x , y ∈ X × Y

where X is the input and Y is the output of the channel.

Example

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 4 / 29

Page 6: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Outline

1 Definitions and TerminologyDiscrete Memoryless ChannelsTerminologyJointly Typical Sets

2 Noisy-Channel Coding TheoremStatementPart onePart twoPart three

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 5 / 29

Page 7: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Definitions

Block Code

A Block Code converts a sequence of source bits s with length K into asequence t of length N with N > K .

Probability of Block Error

The pB of a code and decoder is:∑sin

P(sin)P(sout 6= sin|sin).

Optimal Decoder

An optimal decoder is the decoder which minimalises the probability ofblock error, by decoding an output y as input s, where P(s|y) ismaximalised.

Probability of Bit Error

The pb of a code and decoder is the average probability that a bit is sout isnot equal to the correspoding bit in sin.

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 6 / 29

Page 8: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Outline

1 Definitions and TerminologyDiscrete Memoryless ChannelsTerminologyJointly Typical Sets

2 Noisy-Channel Coding TheoremStatementPart onePart twoPart three

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 7 / 29

Page 9: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Typical Sequences

Definition

Let X be a random variable over an alphabet X . A sequence x ∈ XN oflength N is called typical to tolerance β if and only if

| 1N· log

1

P(x)− H(X )| < β

Example

Suppose X is the result of a coin flip. The sequence

x := 1000111001101100

is typical to any tolerance β ≥ 0.

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 8 / 29

Page 10: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Jointly Typical Sequences

Definition

Let X ,Y be random variables over alphabets X and Y. Two sequencesx ∈ XN and y ∈ Y of length N are called jointly typical to tolerance β ifand only if both x and y are typical and

| 1N· log

1

P(x , y)− H(X ,Y )| < β

Example

Suppose X and Y are both random variables such that P(x = 1) = 0.5and P(y |x) corresponds to a channel of noise level 0.3. The sequences

x := 1111100000

y := 0001100000

are typical to any tolerance β ≥ 0.

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29

Page 11: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Jointly Typical Sequences

Definition

Let X ,Y be random variables over alphabets X and Y. Two sequencesx ∈ XN and y ∈ Y of length N are called jointly typical to tolerance β ifand only if both x and y are typical and

| 1N· log

1

P(x , y)− H(X ,Y )| < β

Example

Suppose X and Y are both random variables such that P(x = 1) = 0.5and P(y |x) corresponds to a channel of noise level 0.3. The sequences

x := 1111100000

y := 0001100000

are typical to any tolerance β ≥ 0.

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29

Page 12: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Jointly Typical Set 1

Definition

Let X ,Y be random variables over alphabets X and Y. The set JN,β thatcontains all pairs (x , y) ∈ X × Y of length N jointly typical to tolerance βis called the jointly-typical set.

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 10 / 29

Page 13: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Jointly Typical Theorem

Theorem

Let x , y be drawn from (XY )N which is defined by

P(x , y) =n∏

i=1

P(xn, yn)

1 The probability that x , y are jointly typical to tolerance β tends to 1as N →∞

2 The number of jointly typical sequences |JN,β| ≤ 2N(H(X ,Y )+β)

3 For any two sequences x , y chosen independently from XN and Y N

respectively that have the same marginal distribution as P(x , y) wehave

P((x , y) ∈ JN,β) ≤ 2−N(I (X ,Y )−3β)

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 11 / 29

Page 14: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Jointly Typical Theorem

Theorem

Let x , y be drawn from (XY )N which is defined by

P(x , y) =n∏

i=1

P(xn, yn)

1 The probability that x , y are jointly typical to tolerance β tends to 1as N →∞

2 The number of jointly typical sequences |JN,β| ≤ 2N(H(X ,Y )+β)

3 For any two sequences x , y chosen independently from XN and Y N

respectively that have the same marginal distribution as P(x , y) wehave

P((x , y) ∈ JN,β) ≤ 2−N(I (X ,Y )−3β)

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 11 / 29

Page 15: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Jointly Typical Theorem

Theorem

Let x , y be drawn from (XY )N which is defined by

P(x , y) =n∏

i=1

P(xn, yn)

1 The probability that x , y are jointly typical to tolerance β tends to 1as N →∞

2 The number of jointly typical sequences |JN,β| ≤ 2N(H(X ,Y )+β)

3 For any two sequences x , y chosen independently from XN and Y N

respectively that have the same marginal distribution as P(x , y) wehave

P((x , y) ∈ JN,β) ≤ 2−N(I (X ,Y )−3β)

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 11 / 29

Page 16: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

An Illustration

Figure: Typical Sets (from MacKay 2003)

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 12 / 29

Page 17: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Outline

1 Definitions and TerminologyDiscrete Memoryless ChannelsTerminologyJointly Typical Sets

2 Noisy-Channel Coding TheoremStatementPart onePart twoPart three

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 13 / 29

Page 18: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Statement

Theorem1 For every discrete memoryless channel. the channel capacity

C = maxPX

I (X ;Y )

satisfies the following property. For any ε > 0 And rate R < C , forsufficiently large N, there is a code of length N and rate ≥ R and adecoding algorithm, such that the maximimal probability of blockerror is < ε.

2 If we accept bit error with probability pb, it is possible to achieverates up to R(pb), where

R(pb) =C

1− h(pb).

3 Rates greater than R(pb) are not achievable.

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 14 / 29

Page 19: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Statement

Theorem1 For every discrete memoryless channel. the channel capacity

C = maxPX

I (X ;Y )

satisfies the following property. For any ε > 0 And rate R < C , forsufficiently large N, there is a code of length N and rate ≥ R and adecoding algorithm, such that the maximimal probability of blockerror is < ε.

2 If we accept bit error with probability pb, it is possible to achieverates up to R(pb), where

R(pb) =C

1− h(pb).

3 Rates greater than R(pb) are not achievable.

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 14 / 29

Page 20: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Statement

Theorem1 For every discrete memoryless channel. the channel capacity

C = maxPX

I (X ;Y )

satisfies the following property. For any ε > 0 And rate R < C , forsufficiently large N, there is a code of length N and rate ≥ R and adecoding algorithm, such that the maximimal probability of blockerror is < ε.

2 If we accept bit error with probability pb, it is possible to achieverates up to R(pb), where

R(pb) =C

1− h(pb).

3 Rates greater than R(pb) are not achievable.

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 14 / 29

Page 21: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Outline

1 Definitions and TerminologyDiscrete Memoryless ChannelsTerminologyJointly Typical Sets

2 Noisy-Channel Coding TheoremStatementPart onePart twoPart three

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 15 / 29

Page 22: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Proof: An Analogy

Figure: Weighing Babies (from MacKay 2003)

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 16 / 29

Page 23: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Coding

Creating a Code

Consider a fixed distribution P(x). We will generate S = 2NR′

codewordsat random using

P(x) =N∏

n=1

P(xn)

and assign a codeword x (s) to each message s. We make this code knownto both sides of the channel.

Important!

This code has a rate of R ′!

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 17 / 29

Page 24: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Decoding

Received Signal

The signal received on the other end of the channel is y , with

P(y |x (s)) =N∏

n=1

P(yn|x (s)n )

Decoding

We will decode using typical-set decoding. We will decode y as s if(x (s), y) are jointly typical and there is no other message s ′ such that(x (s

′), y) are jointly typical.

Mistakes

We will make a mistake when

1 There is no jointly typical x (s).

2 There are multiple jointly typical x (s).

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 18 / 29

Page 25: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Three Types of Errors

Block Error

pB(C) ≡ P(s 6= s|C)

Average Block Error

< pB >≡∑C P(s 6= s|C)P(C)

Maximal Block Error

pBM(C) ≡ maxs P(s 6= s|s, C)

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 19 / 29

Page 26: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Three Types of Errors

Block Error

pB(C) ≡ P(s 6= s|C)

Average Block Error

< pB >≡∑C P(s 6= s|C)P(C)

Maximal Block Error

pBM(C) ≡ maxs P(s 6= s|s, C)

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 19 / 29

Page 27: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Three Types of Errors

Block Error

pB(C) ≡ P(s 6= s|C)

Average Block Error

< pB >≡∑C P(s 6= s|C)P(C)

Maximal Block Error

pBM(C) ≡ maxs P(s 6= s|s, C)

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 19 / 29

Page 28: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Bounding the Errors (1)

No Jointly Typical x (s)

By the first part of the jointly typical theorem

∀δ ∃N(δ) : P(x (1), y) 6∈ JN,β) < 2δ

Too Many Jointly Typical x (s)

The chance for a random x (s′) and y to be jointly typical

≤ 2−N(I (X ;Y )−3β. There are 2(NR′) − 1 candidates.

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 20 / 29

Page 29: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Bounding the Errors (2)

Average Block Error

Now, using the union bound we find

< pBM > ≤ δ +2NR

′∑s′=2

2−N(I (X ;Y )−3β

≤ δ + 2−N(I (X ;Y )−R′−3β)

If R ′ < I (X ;Y )− 3β we can make this error very small (smaller than 2δ).

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 21 / 29

Page 30: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Three modifications

Pick the best P(x)

We choose the best P(x), so R ′ < I (X ;Y )− 3β becomes R ′ < C − 3β.

Pick the best CUsing our baby-argument, there must be a code with pB(C) < 2δ

Perform a trick

From this C we now toss the worst half of the codewords. Those thatremain must have probability of error < 4δ. We define a new code withthese (2NR

′−1) codewords.

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 22 / 29

Page 31: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Conclusion

We have proven the existence of a code C with rate R ′ < C − 3β with amaximal probability of error < 4δ. The theorem can now be proven besetting

R ′ =R + C

2

δ =ε

4

β <(C − R ′)

3N big enough

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 23 / 29

Page 32: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Outline

1 Definitions and TerminologyDiscrete Memoryless ChannelsTerminologyJointly Typical Sets

2 Noisy-Channel Coding TheoremStatementPart onePart twoPart three

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 24 / 29

Page 33: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Overview1 What happens when we try to communicate with a rate, greater than

the capacity?

2 We could just send 1/R of the source bits and guess the rest of thesource. This will give us an average pb of 1

2(1− 1/R).

3 However, it turns out we can minimalise this error so that we get:H2(pb) = 1− 1/R.

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 25 / 29

Page 34: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Method1 We take an excellent (N,K ) code with rate R ′ = K/N.

2 This code is capable of correcting errors in our channel with transitionprobability q.

3 Asymptotically we may assume that R ' 1− H2(pb).

4 We know chop our source our source up in blocks of length N andpass it through our decoder, which gives us blocks of length K , whichthen get communicated over the noiseless channel.

5 When we pass this new message to our encoder, we will receiver amessage which we will differ at an average of qN bits from theoriginal message, so pb = q.

6 Attaching this compressor to our capacity-C error-free communicatorwe get a rate of R = NC

K = CR′ = C

1−H2(pb).

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 26 / 29

Page 35: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Outline

1 Definitions and TerminologyDiscrete Memoryless ChannelsTerminologyJointly Typical Sets

2 Noisy-Channel Coding TheoremStatementPart onePart twoPart three

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 27 / 29

Page 36: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Proof1 A Markov chain is defined by the source, encoder, noisy channel and

the decoder:

P(s, x , y , s) = P(s)P(x |s)P(y |x)P(s|y)

2 I (s; s) ≤ I (x ; y) because of the data processing inequality.

3 By the definition of channel capacity we know that I (x ; y) ≤ NC , soI (s; s) ≤ NC .

4 Assuming that our system has a rate R and bit error probability pb,then I (s; s) ≥ NR(1− H2(pb)).

5 If R > R(pb) = C1−H2(pb)

, then I (s; s) ≥ NC . This gives acontradiction, so for any pb, there is no larger rate possible thanR(pb).

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 28 / 29

Page 37: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Proof1 A Markov chain is defined by the source, encoder, noisy channel and

the decoder:

P(s, x , y , s) = P(s)P(x |s)P(y |x)P(s|y)

2 I (s; s) ≤ I (x ; y) because of the data processing inequality.

3 By the definition of channel capacity we know that I (x ; y) ≤ NC , soI (s; s) ≤ NC .

4 Assuming that our system has a rate R and bit error probability pb,then I (s; s) ≥ NR(1− H2(pb)).

5 If R > R(pb) = C1−H2(pb)

, then I (s; s) ≥ NC . This gives acontradiction, so for any pb, there is no larger rate possible thanR(pb).

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 28 / 29

Page 38: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Proof1 A Markov chain is defined by the source, encoder, noisy channel and

the decoder:

P(s, x , y , s) = P(s)P(x |s)P(y |x)P(s|y)

2 I (s; s) ≤ I (x ; y) because of the data processing inequality.

3 By the definition of channel capacity we know that I (x ; y) ≤ NC , soI (s; s) ≤ NC .

4 Assuming that our system has a rate R and bit error probability pb,then I (s; s) ≥ NR(1− H2(pb)).

5 If R > R(pb) = C1−H2(pb)

, then I (s; s) ≥ NC . This gives acontradiction, so for any pb, there is no larger rate possible thanR(pb).

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 28 / 29

Page 39: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Proof1 A Markov chain is defined by the source, encoder, noisy channel and

the decoder:

P(s, x , y , s) = P(s)P(x |s)P(y |x)P(s|y)

2 I (s; s) ≤ I (x ; y) because of the data processing inequality.

3 By the definition of channel capacity we know that I (x ; y) ≤ NC , soI (s; s) ≤ NC .

4 Assuming that our system has a rate R and bit error probability pb,then I (s; s) ≥ NR(1− H2(pb)).

5 If R > R(pb) = C1−H2(pb)

, then I (s; s) ≥ NC . This gives acontradiction, so for any pb, there is no larger rate possible thanR(pb).

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 28 / 29

Page 40: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

Proof1 A Markov chain is defined by the source, encoder, noisy channel and

the decoder:

P(s, x , y , s) = P(s)P(x |s)P(y |x)P(s|y)

2 I (s; s) ≤ I (x ; y) because of the data processing inequality.

3 By the definition of channel capacity we know that I (x ; y) ≤ NC , soI (s; s) ≤ NC .

4 Assuming that our system has a rate R and bit error probability pb,then I (s; s) ≥ NR(1− H2(pb)).

5 If R > R(pb) = C1−H2(pb)

, then I (s; s) ≥ NC . This gives acontradiction, so for any pb, there is no larger rate possible thanR(pb).

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 28 / 29

Page 41: Shannon's Noisy-Channel Coding Theoremschaffne/courses/infcom/... · Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 9 / 29. Jointly Typical Sequences

References

David J.C. MacKay - Information Theory, Inference, and LearningAlgorithms - Cambridge University Press 2003 - Accessed viahttp://www.inference.phy.cam.ac.uk/itprnn/book.pdf

Lucas Slot, Sebastian Zur Shannon’s Noisy-Channel Coding Theorem February 13, 2015 29 / 29