Exercise 1. = P(y a 1)P(a 1 )

Similar documents
Shannon s noisy-channel theorem

Capacity of a channel Shannon s second theorem. Information Theory 1/33

X 1 : X Table 1: Y = X X 2

ECE Information theory Final

Notes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel

National University of Singapore Department of Electrical & Computer Engineering. Examination for

16.36 Communication Systems Engineering

EE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15

Noisy channel communication

Coding of memoryless sources 1/35

CSCI 2570 Introduction to Nanocomputing

Error Detection and Correction: Hamming Code; Reed-Muller Code

Physical Layer and Coding

Lecture 3: Channel Capacity

Entropies & Information Theory

Digital Communications III (ECE 154C) Introduction to Coding and Information Theory

Lecture 3 : Algorithms for source coding. September 30, 2016

Lecture 15: Conditional and Joint Typicaility

Shannon s Noisy-Channel Coding Theorem

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Lecture 10: Broadcast Channel and Superposition Coding

EE376A - Information Theory Final, Monday March 14th 2016 Solutions. Please start answering each question on a new page of the answer booklet.

channel of communication noise Each codeword has length 2, and all digits are either 0 or 1. Such codes are called Binary Codes.

Lecture 8: Shannon s Noise Models

Channel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5.

9 THEORY OF CODES. 9.0 Introduction. 9.1 Noise

ECE Information theory Final (Fall 2008)

Noisy-Channel Coding

Lecture 18: Shanon s Channel Coding Theorem. Lecture 18: Shanon s Channel Coding Theorem

1 Introduction to information theory

Electrical and Information Technology. Information Theory. Problems and Solutions. Contents. Problems... 1 Solutions...7

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

Appendix B Information theory from first principles

Lecture 8: Channel and source-channel coding theorems; BEC & linear codes. 1 Intuitive justification for upper bound on channel capacity

EE5585 Data Compression May 2, Lecture 27

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information

Lecture 4 Noisy Channel Coding

Distributed Source Coding Using LDPC Codes

18.2 Continuous Alphabet (discrete-time, memoryless) Channel

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Dept. of Linguistics, Indiana University Fall 2015

Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels

Ma/CS 6b Class 25: Error Correcting Codes 2

(each row defines a probability distribution). Given n-strings x X n, y Y n we can use the absence of memory in the channel to compute

Lecture 7. Union bound for reducing M-ary to binary hypothesis testing

EE 4TM4: Digital Communications II. Channel Capacity

Principles of Communications

Lecture 8: Channel Capacity, Continuous Random Variables

4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information

And for polynomials with coefficients in F 2 = Z/2 Euclidean algorithm for gcd s Concept of equality mod M(x) Extended Euclid for inverses mod M(x)

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

Lecture 5: Channel Capacity. Copyright G. Caire (Sample Lectures) 122

3F1 Information Theory, Lecture 3

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions

ECE Advanced Communication Theory, Spring 2009 Homework #1 (INCOMPLETE)

Solutions to Homework Set #1 Sanov s Theorem, Rate distortion

MATH Examination for the Module MATH-3152 (May 2009) Coding Theory. Time allowed: 2 hours. S = q

COMPSCI 650 Applied Information Theory Apr 5, Lecture 18. Instructor: Arya Mazumdar Scribe: Hamed Zamani, Hadi Zolfaghari, Fatemeh Rezaei

Lecture 4: Codes based on Concatenation

Information Theory CHAPTER. 5.1 Introduction. 5.2 Entropy

Information Theory - Entropy. Figure 3

Solutions to Homework Set #2 Broadcast channel, degraded message set, Csiszar Sum Equality

Coding for Discrete Source

An introduction to basic information theory. Hampus Wessman

Module 1. Introduction to Digital Communications and Information Theory. Version 2 ECE IIT, Kharagpur

MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006)

Introduction to Information Theory. By Prof. S.J. Soni Asst. Professor, CE Department, SPCE, Visnagar

Lecture 6: Expander Codes

Solutions to Homework Set #3 Channel and Source coding

Lecture 4 Channel Coding

Lecture 2. Capacity of the Gaussian channel

Optimum Soft Decision Decoding of Linear Block Codes

1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H.


MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK. SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A

ELEMENT OF INFORMATION THEORY

3F1 Information Theory, Lecture 3

Linear Algebra. F n = {all vectors of dimension n over field F} Linear algebra is about vectors. Concretely, vectors look like this:

ELEC546 Review of Information Theory

Exercises with solutions (Set B)

Mismatched Multi-letter Successive Decoding for the Multiple-Access Channel

Chapter 3 Source Coding. 3.1 An Introduction to Source Coding 3.2 Optimal Source Codes 3.3 Shannon-Fano Code 3.4 Huffman Code

Lecture 12. Block Diagram

Variable Length Codes for Degraded Broadcast Channels

exercise in the previous class (1)

Chapter 9 Fundamental Limits in Information Theory

Lecture 2: August 31

LECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem

Bandwidth: Communicate large complex & highly detailed 3D models through lowbandwidth connection (e.g. VRML over the Internet)

List Decoding of Reed Solomon Codes

6.02 Fall 2011 Lecture #9

Quiz 2 Date: Monday, November 21, 2016

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Discussion 6A Solution

Data Compression Techniques

Digital communication system. Shannon s separation principle

Exam 1. Problem 1: True or false

Lecture 1: Shannon s Theorem

On the Capacity of the Two-Hop Half-Duplex Relay Channel

The Gallager Converse

Transcription:

Chapter 7 Channel Capacity Exercise 1 A source produces independent, equally probable symbols from an alphabet {a 1, a 2 } at a rate of one symbol every 3 seconds. These symbols are transmitted over a binary symmetric channel (with error probability p < 1/2) which is used once per second by an encoding the source symbol a 1 as 000 and the source symbol a 2 as 111. If in the corresponding 3 second interval of the channel output, any of the sequences 000, 001, 010, 100 is received, a 1 is decoded. otherwise, a 2 is decoded. 1. For each possible received 3bit sequence find the probability that a 1 came out of source given that received sequence. 2. Find the probability of an incorrect decision (using question 1. is not the easiest way). 3. Show that the above decoding rule minimizes the probability of an incorrect decision. Answer of exercise 1 1. Let us assume the sequence y was received. Let 0 y denotes the number of 0 in y and 1 y analogously for 1. We want to determine the probability P(a 1 sent y received), shortly P(a 1 y). It is easy to see that P(a 1 y) = P(a 1, y) P(y) = P(y a 1)P(a 1 ). P(y) We know that P(a 1 ) = 1/2, and P(y a 1 ) is uniquely determined by the number of bits that were flipped. Since a 1 is encoded as 000, the number of bits the channel flips when transforming 000 to y is determined by the Hamming weight of y. We get P(y a 1 ) = p 1y (1 p) 0y. 1

2 CHAPTER 7. CHANNEL CAPACITY It remains to determine P(y): Finally, P(y) =P(y a 1 )P(a 1 ) + P(y a 2 )P(a 2 ) P(a 1 y) = = 1 2 (p 1y (1 p) 0y + p 0y (1 p) 1y ) p 1y (1 p) 0y p 1y (1 p) 0y + p 0y (1 p) 1y. 2. We calculate the respective error probabilities provided the signal a 1 (a 2 ) was sent λ 1 = P(y a 1 ) = p 3 + 3p 2 (1 p) λ 2 = y {011,101,110,111} y {100,010,001,000} P(y a 2 ) = p 3 + 3p 2 (1 p). Since λ 1 = λ 2, the (arithmetic) average error probability P e (3) and imum error probability λ coincide. 3. We have only two codewords, so the probability that a sequence y decoded as (WLOG) a 1 was decoded incorrectly is proportional to the number of flips needed to map 111 to y and reads P e a 1 (y) def = p 0y (1 p) 1y. To minimize error probability for particular y, we have to decide to decode it either as a 1 with the aforementioned error probability or as a 2 with the error probability P e a 1 (y) = p 0y (1 p) 1y = P(y a 2 ). P e a 2 (y) = p 1y (1 p) 0y = P(y a 1 ). We observe that the (arithmetic) average error probability is the sum P e (3) = 1 p y, (7.1) 2 y {0,1} 3 where p y is, depending on the decoding rule, either P(y a 1 ) or P(y a 2 ). To minimize the expression (7.1), we have to simply minimize all summands, i.e. for each y set p y = min{p(y a 1 ), P(y a 2 )}. Recall that p < 1/2, what gives that the optimum has to imize the power of p. Hence, our decoding is minimal, observing that p 2 (1 p) < p(1 p) 2 and p 3 < (1 p) 3. We have that for each fixed decoding λ P e (3). Our decoding minimizes P e (3) (3) to be P e. So, the imal error probability for our decoding is λ = P (3) e < P (3) e λ. (7.2)

Exercise 2 Consider a binary symmetric channel with Y i = Z i i, where i, Y i, Z i {0, 1}. Suppose that all Z i have the same marginal probability distribution P (Z i = 1) = p = 1 P (Z i = 0), but that Z 1, Z 2,..., Z n are not necessarily independent. Assume that Z n is independent of n (random vectors are independent!). Let C = 1 H(p, 1 p) = 1 H(Z 1 ). Show that I( 1, 2,..., n ; Y 1, Y 2,..., Y n ) nc. 1, 2,..., n 3 We evaluate Answer of exercise 2 I( 1, 2,..., n ; Y 1, Y 2,..., Y n ) =H(Y 1, Y 2,..., Y n ) H(Y 1, Y 2,..., Y n 1, 2,..., n ) =H(Y 1, Y 2,..., Y n ) H(Z 1, Z 2,..., Z n ). (7.3) The last expression is imized when the joint distribution of Y 1, Y 2,..., Y n is uniform, what is achieved regardless of distribution on Z 1, Z 2,..., Z n via the uniform distribution on 1, 2,..., n, since n and Z n are independent. Using the imum value H(Y 1, Y 2,..., Y n ) = n we get the required result I( 1, 2,..., n ; Y 1, Y 2,..., Y n ) =n H(Z 1, Z 2,..., Z n ) 1, 2,..., n n nh(z 1 ) = nc. (7.4) Exercise 3 Consider the discrete memoryless channel Y = + Z mod 11, where Z is given by i 1 2 3 P(Z = i) 1/3 1/3 1/3 and {0, 1,..., 10}. Calculate the capacity of this channel assuming that Z is independent of. We calculate Answer of exercise 3 ( ) I(; Y ) = (H(Y ) H(Y )) = H(Y ) log 3, (7.5) where ( ) holds because the entropy H(Y = x) does not depend on x, giving it does not depend on the distribution of, and equals to log 3. It remains to imize H(Y ), but it is easy to see that we can achieve the uniform distribution on Y.

4 CHAPTER 7. CHANNEL CAPACITY The imum value is achieved for the uniform distribution of. I(; Y ) = log 10 log 3 (7.6) Exercise 4 Consider two discrete memoryless channels (with independent noise) ( 1, p(y 1 x 1 ), Y 1 ) and ( 2, p(y 2 x 2 ), Y 2 ) with the respective capacities C 1 and C 2. We form a new channel ( 1 2, p((y 1, y 2 ) (x 1, x 2 )), Y 1 Y 2 ), where we send (x 1, x 2 ) simultaneously through the previously defined channels to obtain (y 1, y 2 ). The transition probability is given by p((y 1, y 2 ) (x 1, x 2 )) = p(y 1 x 1 )p(y 2 x 2 ). (7.7) Find the capacity of the combined channel. Answer of exercise 4 Note that the imum here should be taken over all joint (!) distributions of 1 and 2. We start with I( 1, 2 ; Y 1, Y 2 ) = (H(Y 1, Y 2 ) H(Y 1, Y 2 1, 2 )). 1, 2 1, 2 Let us evaluate H(Y 1, Y 2 1, 2 ) = p(x 1, x 2, y 1, y 2 ) log p(x 1, x 2 y 1, y 2 ) x 1,x 2,y 1,y 2 ( ) = p(x 1, x 2, y 1, y 2 ) log(p(y 1 x 1 )p(y 2 x 2 )) x 1,x 2,y 1,y 2 = p(x 1, x 2, y 1, y 2 ) log p(y 1 x 1 ) p(x 1, x 2, y 1, y 2 )p(y 2 x 2 )) x 1,x 2,y 1,y 2 x 1,x 2,y 1,y 2 = x 1,y 1 p(x 1, y 1 ) log p(y 1 x 1 ) x 2,y 2 p(x 2, y 2 )p(y 2 x 2 )) =H(Y 1 1 ) + H(Y 2 2 ) using Equation (7.7) to get ( ). We proceed with the capacity I( 1, 2 ; Y 1, Y 2 ) = (H(Y 1, Y 2 ) H(Y 1, Y 2 1, 2 )) 1, 2 1, 2 = 1, 2 (H(Y 1, Y 2 ) H(Y 1 1 ) H(Y 2 2 )) 1, 2 (H(Y 1 ) + H(Y 2 ) H(Y 1 1 ) H(Y 2 2 )) = (H(Y 1 ) H(Y 1 1 )) + (H(Y 2 ) H(Y 2 2 )) 1 2 =C 1 + C 2.

5 On the other hand, we may achieve the capacity C 1 +C 2, since if we imize only over marginal distributions of 1 and 2, we get 1 2 I( 1, 2 ; Y 1, Y 2 ) = (H(Y 1, Y 2 ) H(Y 1, Y 2 1, 2 )) 1 2 = (H(Y 1 ) + H(Y 2 ) H(Y 1 1 ) H(Y 2 2 )) 1 2 = (H(Y 1 ) H(Y 1 1 )) + (H(Y 2 ) H(Y 2 2 )) 1 2 This gives the result =C 1 + C 2. I( 1, 2 ; Y 1, Y 2 ) = C 1 + C 2. (7.8) 1, 2 Consider a 26 key typewriter. Exercise 5 If pushing a key results in printing the associated letter (no errors), what is the capacity C in bits? Suppose that pushing a key results in printing that letter of the next letter (with equal probabilities). Thus, A A or B, B B or C,..., Z Z or A. What is the capacity? What is the highest rate code with block length one that you can find that achieves zero probability of error for the channel in part 2??? Answer of exercise 5 Exercise 6 Let the channel has binary input and output alphabet and transition probabilities p(y x) given by \Y 0 1 0 1 0 1 1/2 1/2 Find capacity of this channel and the input probability distribution achieving this capacity.?? Answer of exercise 6