LECTURE 13. Last time: Lecture outline

Similar documents
LECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem

LECTURE 10. Last time: Lecture outline

Network coding for multicast relation to compression and generalization of Slepian-Wolf

LECTURE 2. Convexity and related notions. Last time: mutual information: definitions and properties. Lecture outline

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions

LECTURE 3. Last time:

Lecture 3: Channel Capacity

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

Lecture 8: Channel and source-channel coding theorems; BEC & linear codes. 1 Intuitive justification for upper bound on channel capacity

EE 4TM4: Digital Communications II. Channel Capacity

National University of Singapore Department of Electrical & Computer Engineering. Examination for

Lecture 5: Channel Capacity. Copyright G. Caire (Sample Lectures) 122

Homework Set #2 Data Compression, Huffman code and AEP

Lecture 4 Noisy Channel Coding

Shannon s Noisy-Channel Coding Theorem

Lecture 22: Final Review

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Lecture 1: The Multiple Access Channel. Copyright G. Caire 12

(Classical) Information Theory III: Noisy channel coding

Lecture 4 Channel Coding

Lecture 11: Quantum Information III - Source Coding

4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information

Capacity of a channel Shannon s second theorem. Information Theory 1/33

Shannon s noisy-channel theorem

ELEC546 Review of Information Theory

Lecture 14 February 28

Lecture 2. Capacity of the Gaussian channel

ECE 4400:693 - Information Theory

The Method of Types and Its Application to Information Hiding

Communications Theory and Engineering

Information Theory. Lecture 10. Network Information Theory (CT15); a focus on channel capacity results

Information Theory for Wireless Communications. Lecture 10 Discrete Memoryless Multiple Access Channel (DM-MAC): The Converse Theorem

Lecture 10: Broadcast Channel and Superposition Coding

X 1 : X Table 1: Y = X X 2

Lecture 15: Conditional and Joint Typicaility

for some error exponent E( R) as a function R,

EE376A - Information Theory Final, Monday March 14th 2016 Solutions. Please start answering each question on a new page of the answer booklet.

An Extended Fano s Inequality for the Finite Blocklength Coding

Chapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye

Solutions to Set #2 Data Compression, Huffman code and AEP

EE376A: Homeworks #4 Solutions Due on Thursday, February 22, 2018 Please submit on Gradescope. Start every question on a new page.

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria

Lecture 5 Channel Coding over Continuous Channels

MARKOV CHAINS A finite state Markov chain is a sequence of discrete cv s from a finite alphabet where is a pmf on and for

On the Duality between Multiple-Access Codes and Computation Codes

The Capacity Region for Multi-source Multi-sink Network Coding

Lecture 5: Asymptotic Equipartition Property

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018

EE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15

Upper Bounds on the Capacity of Binary Intermittent Communication

On Function Computation with Privacy and Secrecy Constraints

Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels

An introduction to basic information theory. Hampus Wessman

On Multiple User Channels with State Information at the Transmitters

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16

Chapter 9. Gaussian Channel

Lecture 11: Polar codes construction

16.36 Communication Systems Engineering

The Communication Complexity of Correlation. Prahladh Harsha Rahul Jain David McAllester Jaikumar Radhakrishnan

Channel Coding for Secure Transmissions

Multicoding Schemes for Interference Channels

EE5585 Data Compression May 2, Lecture 27

18.2 Continuous Alphabet (discrete-time, memoryless) Channel

A Novel Asynchronous Communication Paradigm: Detection, Isolation, and Coding

EE229B - Final Project. Capacity-Approaching Low-Density Parity-Check Codes

Capacity bounds for multiple access-cognitive interference channel

Cut-Set Bound and Dependence Balance Bound

Massachusetts Institute of Technology. Problem Set 4 Solutions

Computing and Communications 2. Information Theory -Entropy

Network Coding on Directed Acyclic Graphs

Diversity-Multiplexing Tradeoff in MIMO Channels with Partial CSIT. ECE 559 Presentation Hoa Pham Dec 3, 2007

Lecture 17: Differential Entropy

On the Capacity Region for Secure Index Coding

EE5319R: Problem Set 3 Assigned: 24/08/16, Due: 31/08/16

On the Capacity Region of the Gaussian Z-channel

(each row defines a probability distribution). Given n-strings x X n, y Y n we can use the absence of memory in the channel to compute

Relay Networks With Delays

Lecture 7 September 24

EE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm

Capacity of AWGN channels

Frans M.J. Willems. Authentication Based on Secret-Key Generation. Frans M.J. Willems. (joint work w. Tanya Ignatenko)

ECE Advanced Communication Theory, Spring 2009 Homework #1 (INCOMPLETE)

A Graph-based Framework for Transmission of Correlated Sources over Multiple Access Channels

Secret Key Agreement Using Asymmetry in Channel State Knowledge

The Poisson Channel with Side Information

Lecture Notes for Communication Theory

Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets

On the Secrecy Capacity of Fading Channels

Performance-based Security for Encoding of Information Signals. FA ( ) Paul Cuff (Princeton University)

5 Mutual Information and Channel Capacity

Information Theory. M1 Informatique (parcours recherche et innovation) Aline Roumy. January INRIA Rennes 1/ 73

Lecture Notes on Digital Transmission Source and Channel Coding. José Manuel Bioucas Dias

Generalized Writing on Dirty Paper

Appendix B Information theory from first principles

Two Applications of the Gaussian Poincaré Inequality in the Shannon Theory

Shannon s Noisy-Channel Coding Theorem

Approaching Blokh-Zyablov Error Exponent with Linear-Time Encodable/Decodable Codes

On the Capacity and Degrees of Freedom Regions of MIMO Interference Channels with Limited Receiver Cooperation

Solutions to Homework Set #1 Sanov s Theorem, Rate distortion

On Achievable Rates for Channels with. Mismatched Decoding

Transcription:

LECTURE 13 Last time: Strong coding theorem Revisiting channel and codes Bound on probability of error Error exponent Lecture outline Fano s Lemma revisited Fano s inequality for codewords Converse to the coding theorem Reading: Sct. 8.9.

Fano s lemma Suppose we have r.v.s X and Y, Fano s lemma bounds the error we expect when estimating X from Y We generate an estimator of X that is X = g(y ). Probability of error P e = P r( X = X) Indicator function for error E which is 0 when X = X and 1 otherwise. Thus, P e = P (E = 1) Fano s lemma: H(E) + P e log( X 1) H(X Y ) We now need to consider the case where we are dealing with codewords Want to show that vanishingly small probability of error is not possible if the rate is above capacity

Fano s inequality for code words An error occurs when the decoder makes the wrong decision in selecting the message that was transmitted Let M {1, 2,..., 2 nr } be the transmitted message and let M be the estimate of the received message from Y n M is uniformly distributed in {1, 2,..., 2 nr } and consecutive message transmissions are IID (thus, we do not make use of a number of messages, but consider a single message transmission) The probability of error for a codebook for transmission of Mis P e,m = P (M = M) = E Y n[p (M = M Y n )] Consider an indicator variable E = 1 when an error occurs and E = 0 otherwise

Fano s inequality for code words H(E, M Y ) = H(M Y ) + H(E M, Y ) = H(M Y ) = H(E Y ) + H(M E, Y ) 1 + H(M E, Y ) Let us consider upper bounding the RHS H(M E, Y ) we are not averaging over codebooks as for the coding theorem, but are considering a specific codebook = H(X E, Y ) = E M,Y [P (M = M Y )]H(X E = 1, Y ) + (1 E M,Y [P (M = M Y )]) H(X E = 0, Y ) = P e H(X E = 1, Y ) P e H(X E = 1) P e log( M 1)

Fano s inequality for code words Given the definition of rate, M = 2 nr, so H(M E, Y ) P e nr + 1 Hence H(M Y ) P e nr For a given codebook, M determines X, so H(X Y ) = H(M Y ) 1 + P e nr for a DMC with a given codebook and uniformly distributed input messages

From Fano s inequality for code words to the coding theorem converse We now want to relate this to mutual information and to capacity Strategy: - will need to have mutual information expressed as H(M) H(M Y ) - rate will need to come in play - try the fact that H(M) = nr for uniformly distributed messages - will need capacity to come into play. We remember that combining the chain rule for entropies and the fact that conditioning reduces entropy yields the fact that for a DMC I(X n ; Y n ) nc

Converse to the channel coding theorem Consider some sequence of codebooks (2 nr, n), indexed by n, such that the maximum probability of error over each codebook goes to 0 as n goes to Assume (we ll revisit this later) that the message M is drawn with uniform PMF from {1, 2,..., 2 nr } Then nr = H(M) Also H(M) = H(M Y ) + I(M; Y ) = H(M Y ) + H(Y ) H(Y M) = H(M Y ) + H(Y ) H(Y X) = H(M Y ) + I(X; Y ) 1 + P e nr + nc Hence R 1 + P e R + C n

Converse to the channel coding theorem Letting n go to, we obtain that R C (since the maximum probability of error goes to 0 by our assumption) Moreover, we obtain the following bound on error: P e 1 C R nr 1 Note: - for R < C, the bound has a negative RHS, so does not bound probability of error in a way that is inconsistent with forward coding theorem - for R > C, bound becomes 1 C for large R n, but 1 C 1 is always lower bound R R - as R goes to infinity, bound becomes 1, so is tight bound - RHS of bound does not vary with n in the way we would expect, since the bound increases with n

Revisiting the message distribution We have assumed that we can select the messages to be uniformly distributed This is crucial to get H(M) = nr Does the converse only work when the messages are uniformly distributed? Let us revisit the consequences of the AEP

Consequences of the AEP: the typical set (n) Definition: A ɛ is a typical set with respect to P X (x) if it is the set of sequences in the set of all possible sequences x n X n with probability: 2 n(h(x)+ɛ) P X n (x n ) 2 n(h(x) ɛ) equivalently 1 H(X) ɛ n log(p X n (x n )) H(X) + ɛ We shall use the typical set to describe a set with characteristics that belong to the majority of elements in that set.

Consequences of the AEP: the typical set Why is it typical? The probability of being more than δ away from H(X) goes can be arbitrarily close to 0 as n, hence (n) P r(a ɛ ) 1 ɛ We can select ɛ to be arbitrarily small, so that the distribution of messages is arbitrarily close to uniform in the typical set The max of the probability of error must be bounded away from 0 in the typical set for the max of the probability of error to be bounded away from 0 The probability of error is dominated by the probability of the typical set as we let ɛ > 0

MIT OpenCourseWare http://ocw.mit.edu 6.441 Information Theory Spring 2010 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.