Lecture 18: Shanon s Channel Coding Theorem. Lecture 18: Shanon s Channel Coding Theorem

Similar documents
Notes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel

Lecture 4: Codes based on Concatenation

Lecture 8: Shannon s Noise Models

Lecture 4: Proof of Shannon s theorem and an explicit code

Exercise 1. = P(y a 1)P(a 1 )

for some error exponent E( R) as a function R,

Shannon s noisy-channel theorem

CSCI 2570 Introduction to Nanocomputing

Lecture 4 Channel Coding

atri/courses/coding-theory/book/

MATH Examination for the Module MATH-3152 (May 2009) Coding Theory. Time allowed: 2 hours. S = q

Error Correcting Codes: Combinatorics, Algorithms and Applications Spring Homework Due Monday March 23, 2009 in class

Lecture 4 Noisy Channel Coding

(each row defines a probability distribution). Given n-strings x X n, y Y n we can use the absence of memory in the channel to compute

Lecture B04 : Linear codes and singleton bound

Lecture 11: Polar codes construction

Shannon s Noisy-Channel Coding Theorem

Capacity of a channel Shannon s second theorem. Information Theory 1/33

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018

LECTURE 10. Last time: Lecture outline

Lecture 3: Channel Capacity

Introduction to Convolutional Codes, Part 1

List Decoding of Reed Solomon Codes

Lecture 8: Channel and source-channel coding theorems; BEC & linear codes. 1 Intuitive justification for upper bound on channel capacity

Limits to List Decoding Random Codes

Ma/CS 6b Class 25: Error Correcting Codes 2

National University of Singapore Department of Electrical & Computer Engineering. Examination for

EE376A - Information Theory Final, Monday March 14th 2016 Solutions. Please start answering each question on a new page of the answer booklet.

EE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15

Lecture 17: Perfect Codes and Gilbert-Varshamov Bound

Entropies & Information Theory

Entropy. Probability and Computing. Presentation 22. Probability and Computing Presentation 22 Entropy 1/39

X 1 : X Table 1: Y = X X 2

1 Introduction to information theory

Lecture 19 : Reed-Muller, Concatenation Codes & Decoding problem

Lecture 3: Error Correcting Codes

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

An introduction to basic information theory. Hampus Wessman

Computing and Communications 2. Information Theory -Entropy

Lecture 7. Union bound for reducing M-ary to binary hypothesis testing

ELEC546 Review of Information Theory

CS151 Complexity Theory. Lecture 9 May 1, 2017

Coding into a source: an inverse rate-distortion theorem

The first bound is the strongest, the other two bounds are often easier to state and compute. Proof: Applying Markov's inequality, for any >0 we have

Upper Bounds on the Capacity of Binary Intermittent Communication

Belief-Propagation Decoding of LDPC Codes

Low-Density Parity-Check Codes

Lecture 7 September 24

1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H.

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Lecture 15: Conditional and Joint Typicaility

16.36 Communication Systems Engineering

Noisy-Channel Coding

Delays and the Capacity of Continuous-time Channels

Chapter 3 Source Coding. 3.1 An Introduction to Source Coding 3.2 Optimal Source Codes 3.3 Shannon-Fano Code 3.4 Huffman Code

Lecture 12. Block Diagram

Shannon's Theory of Communication

CS 630 Basic Probability and Information Theory. Tim Campbell

6.02 Fall 2011 Lecture #9

Second-Order Asymptotics in Information Theory

Phase Transitions of Random Codes and GV-Bounds

ECEN 655: Advanced Channel Coding

Exponential Error Bounds for Block Concatenated Codes with Tail Biting Trellis Inner Codes

Example 1. The sample space of an experiment where we flip a pair of coins is denoted by:

Bridging Shannon and Hamming: List Error-Correction with Optimal Rate

Lecture 12: Reed-Solomon Codes

Quiz 2 Date: Monday, November 21, 2016

Appendix B Information theory from first principles

Homework Set #2 Data Compression, Huffman code and AEP

Solutions to Set #2 Data Compression, Huffman code and AEP

Shannon s Noisy-Channel Coding Theorem

Lecture 9 Polar Coding

(Classical) Information Theory III: Noisy channel coding

FRAMES IN QUANTUM AND CLASSICAL INFORMATION THEORY

Optimum Soft Decision Decoding of Linear Block Codes

Maximum Likelihood Decoding of Codes on the Asymmetric Z-channel

Error-Correcting Codes:

Lecture 19: Elias-Bassalygo Bound

Towards Optimal Deterministic Coding for Interactive Communication

Intro to Information Theory

Chapter 9 Fundamental Limits in Information Theory

Codes on graphs and iterative decoding

Solutions to Homework Set #3 Channel and Source coding

Information measures in simple coding problems

6.02 Fall 2012 Lecture #1

Lecture 5: Channel Capacity. Copyright G. Caire (Sample Lectures) 122

4 An Introduction to Channel Coding and Decoding over BSC

Random Variable. Pr(X = a) = Pr(s)

Tail Inequalities. The Chernoff bound works for random variables that are a sum of indicator variables with the same distribution (Bernoulli trials).

Entropy as a measure of surprise

ECE Advanced Communication Theory, Spring 2009 Homework #1 (INCOMPLETE)

CSCI-B609: A Theorist s Toolkit, Fall 2016 Oct 4. Theorem 1. A non-zero, univariate polynomial with degree d has at most d roots.

Machine Learning. Computational Learning Theory. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012

Lecture 1: Introduction, Entropy and ML estimation

Coding for a Non-symmetric Ternary Channel

1 Randomized Computation

3F1 Information Theory, Lecture 3

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye

Transcription:

Channel Definition (Channel) A channel is defined by Λ = (X, Y, Π), where X is the set of input alphabets, Y is the set of output alphabets and Π is the transition probability of obtaining a symbol y Y if the input symbol is x X. For example: A Binary Symmetric Channel with flipping probability p (i.e., p-bsc) is a channel with X = {0, 1} and Y = {0, 1}, and the probability of obtaining b given input symbol b is (1 p) and the probability of obtaining (1 b) given input symbol b is p

Channel Capacity Definition (Capacity) The capacity of a channel is defined by: C(Λ) = max H(Y ) H(Y X ) Dist p over X Note that it is not necessary that the maximization happens when p is the uniform distribution over X For p-bsc, the maximization happens for p = U X and the capacity is 1 h(p)

Shannon s Channel Coding Theorem Theorem (Shanon s Channel Coding Theorem) For every channel Λ, there exists a constant C = C(Λ), such that for all 0 R < C, there exists n 0, such that for all n n 0, there exists encoding and decoding algorithms and Dec such that: : {1,..., M = 2 Rn } X n, and PrDec(Π((m))) = m] 1 exp( Ω(n)) English Version: For every channel, there exists a constant capacity, such that for all rate less than the capacity, (for large enough n), we can reliably push information across the channel at that rate

Coding Theorem for BSC Let Z(n, p) be n independent trials of a Bernoulli variable with probability of heads being p Theorem For all p, there exists C = 1 h(p), such that for all 0 R = 1 h(p) ε and ε > 0, there exists n 0, such that for all n n 0, there exists encoding and decoding algorithms and Dec such that: : {0, 1} Rn {0, 1} n, and Pr z Z(n,p) Dec((m) + z) = m] 1 exp( Ω(n)) In fact, there exists a binary linear code that achieves this rate Further, a random binary linear code achieves this rate with probability exponentially close to 1

Proof of Coding Theorem for BSC Define k = (1 h(p) ε)n and l = k + 1 We shall show that there exists an encoding scheme using probabilistic methods Let : {0, 1} l {0, 1} n be a random map Let Dec(y) be the maximum likelihood decoding, i.e. it decodes y to the nearest codeword Fix a message m {0, 1} l We are interested in: Expected (over random ) decoding error probability ] err(m) := E Note that, we have: err(m) E Pr Dec((m) + z) m] z Z(n,p) Pr wt(z) (p + ε)n] z Z(n,p) + Pr wt(z) (p + ε)n Dec((m) + z) m] z Z(n,p) ]

Proof of Coding Theorem for BSC (continued) First error term is at most exp( Ω(ε 2 n)) by Chernoff Bound Let p(z) be the probability of z Z(n, p) 1(E) represents the indicator variable for the event E So, we have: err(m) E err 1 + err 2 (m, )], where: err 1 = exp( Ω(ε 2 n)) err 2 (m, ) = p(z) 1(Dec((m) + z) m) By linearity of expectation, we get: err(m) exp( Ω(ε 2 n)) + E err 2 (m, )]

Proof of Coding Theorem for BSC (continued) We need to bound only: E err 2 (m, )], which turns out to be (by swapping and E operators): = = p(z) E 1(Dec((m) + z) m)] p(z) E m m : 1(Dec((m) + z) = m ) ] p(z) Pr m m : Dec((m) + z) = m ] p(z) Pr m m : c c + Ball 2 (n, (p + ε)n) ] Here c and c, respectively, are (m) and (m )

Proof of Coding Theorem for BSC (continued) By union bound, we get: p(z) p(z) m : m m m : m m p(z) 2 l 2h(p)n 2 n = = 2 2 εn = exp( Ω(n)) Pr c c + Ball 2 (n, (p + ε)n) ] Vol 2 (n, (p + ε)n) 2 n p(z)2 2 εn Overall, we get: For a fixed m, the expected decoding error (over a randomly chosen encoding function) is err(m) E err 1 + err 2 (m, )] exp( Ω(n))

Proof of Coding Theorem for BSC (continued) Therefore, E E m $ {1,...,2 l } = E m $ {1,...,2 l } E m $ {1,...,2 l } E ] Pr Dec((m) + z) m] z Z(n,p) ]] Pr Dec((m) + z) m] z Z(n,p) exp( Ω(n))] = exp( Ω(n)) So, there exists an such that the expected (over random messages) decoding error probability is at most exp( Ω(n)) By pigeon hole principle, for this choice of, at most half the messages in {1,..., 2 l } have decoding error probability 2 exp( Ω(n)) So, for this choice of there exists a subset of {1,..., 2 l } of size 2 k such that each message has decoding error probability 2 exp( Ω(n)) = exp( Ω(n))

Additional Comments If we show that a random linear encoding succeeds then we do not need to perform an averaging over m, because the decoding error probability for a particular m is identical to the decoding error probability for any m (because, in linear codes, the view from a codeword c about the universe is identical to the view of any codeword c about the universe ) We can also show that for 1 exp( Ω(n)) fraction of the decoding error is exponentially small (because, decoding error is bounded by 1 and we can perform an averaging argument)

Converse of Shannon s Channel Coding Theorem Intuitively, Theorem (Converse of Shannon s Channel Coding Theorem) For any channel Λ, there exists C = C(Λ), such that for all R > C, and any encoding and decoding functions and Dec, respectively, (for all n) the decoding error is at least a constant when m $ {1,..., 2 Rn }