Solutions to Homework Set #1 Sanov s Theorem, Rate distortion

Size: px
Start display at page:

Download "Solutions to Homework Set #1 Sanov s Theorem, Rate distortion"

Transcription

1 st Semester 00/ Solutions to Homework Set # Sanov s Theorem, Rate distortion. Sanov s theorem: Prove the simple version of Sanov s theorem for the binary random variables, i.e., let X,X,...,X n be a sequence of binary random variables, drawn i.i.d. according to the distribution: Pr(X = ) = q, Pr(X = 0) = q. () Let the proportion of s in the sequence X,X,...,X n be p X, i.e., p X n = n X i. () i= By the law of large numbers, we would expect p X to be close to q for large n. Sanov s theorem deals with the probability that p X n is far away from q. In particular, for concreteness, if we take p > q >, Sanov s theorem states that n logpr{(x,x,...,x n ) : p X n p} plog p p +( p)log q q = D((p, p) (q, q) (3) Justify the following steps: Pr{(X,X,...,X n ) : p X p} i= np q i ( q) n i (4) i Argue that the term corresponding to i = np is the largest term in the sum on the right hand side of the last equation. Show that this term is approximately nd. Prove an upper bound on the probability in Sanov s theorem using the above steps. Use similar arguments to prove a lower bound and complete the proof of Sanov s theorem.

2 Solution Sanov s theorem Since nx n has a binomial distribution, we have Pr(nX n = i) = q i ( q) n i (5) i and therefore Pr{(X,X,...,X n ) : p X p} i= np q i ( q) n i (6) i Pr(nX n = i+) Pr(nX n = i) = ( n i+( n i ) q i+ ( q) n i ) = n i qi ( q) n i i+ q q (7) This ratio is less than if n i,i.e., if i > nq ( q). Thus q the maximum of the terms occurs when i = np. From Example..3, i+ < q.= nh(p) np and hence the largest term in the sum is q np ( q) n np = n( plogp ( p)log( p))+nplogq+n( p)log( q) = nd(p q) np (9) From the above results, it follows that Pr{(X,X,...,X n ) : p X p} q i ( q) n i (0) i i= np (n np ) q i ( q) () n i np (8) (n( p)+) nd(p q) () where the second inequality follows from the fact that the sum is less than the largest term times the number of terms. Taking the

3 logarithm and dividing by n and taking the limit as n, we obtain lim n n logpr{(x,x,...,x n ) : p X p} D(p q) (3) Similarly, using the fact the sum of the terms is larger than the largest term, we obtain Pr{(X,X,...,X n ) : p X p} q i ( q) n i (4) i i= np q i ( q) n i (5) np and nd(p q) (6) lim n n logpr{(x,x,...,x n ) : p X p} D(p q) (7) Combining these two results, we obtain the special case of Sanov s theorem lim n n logpr{(x,x,...,x n ) : p X p} = D(p q) (8). Strong Typicality Let X n be drawn i.i.d. P(x). Prove that for each x n T δ (X), n(h(x)+δ ) P n (x n ) n(h(x) δ ) for some δ = δ (δ) that vanishes as δ 0. Solution: Strong Typicality From our familiar trick, we have P n (x n ) = n( a X P x n(a)log P(a)) But, since x n T δ (X), we have P x n(a) P(a) < δ for all a. Therefore, H(X) δ log P(a) < P x n(a)log a a X 3 P(a) < H(X)+ δ log a P(a)

4 Hence, where δ = δ a log P(a) n(h(x)+δ ) P n (x n ) n(h(x) δ ) 3. Weak Typicality vs. Strong Typicality In this problem, we compare the weakly typical set A ǫ (P) and the strongly typical set T δ (P). To recall, the definition of two sets are following. { } A ǫ (P) = x n X n : n logpn (x n ) H(P) ǫ { T δ (P) = x n X n : P x n P δ } (a) Suppose P is such that P(a) > 0 for all a X. Then, there is an inclusion relationship between the weakly typical set A ǫ (P) and the strongly typical set T δ (P) for an appropriate choice of ǫ. Which of the statement is true: A ǫ (P) T δ (P) or A ǫ (P) T δ (P)? What is the appropriate relation between δ and ǫ? (b) Give a description of the sequences that belongs to A ǫ (P), vs. the sequences that belongs to T δ (P), when the source is uniformly distributed, i.e. P(a) =, a X. (Assume <.) (c) Can you explain why T δ (P) is called strongly typical set and A ǫ (P) is called weakly typical set? Solution: Weak Typicality vs. Strong Typicality (a) From Problem, we can see that if x n T δ (P), then n logpn (x n ) H(P) = δ log δ P(a) Therefore, we can see that if ǫ > δ a log, then A P(a) ǫ(p) T δ (P). To show that the other way does not hold, see the next part. a 4

5 (b) When P is a uniform distribution, we can see that A ǫ (P) includes every possible sequences x n. To see this, we can easily see that P n (x n ) = ( )n, and thus, /nlogp n (x n ) = log = H(P). Thus, x n A ǫ (P) for all x n, for all ǫ > 0. However, obviously, among those sequences in A ǫ (P), only those who have the type δ P x n(a) + δ for each letter a X, are in T δ (P). Therefore, for sufficiently small value of δ, there always exist some sequences that are in A ǫ (P), but not in T δ (P). (c) From above questions, we can see that the strongly typical set is contained in the weakly typical set for appropriate choice of δ and ǫ. Thus, we can see that the definition of strong typical set is stronger than that of the weakly typical set. 4. Rate distortion for uniform source with Hamming distortion. Consider a source X uniformly distributed on the set {,,...,m}. Find the rate distortion function for this source with Hamming distortion, i.e., { 0 if x = ˆx, d(x,ˆx) = if x ˆx. Solution: Rate distortion for uniform source with Hamming distortion. X is uniformly distributed on the set {,,...,m}. The distortion measure is { 0 if x = ˆx d(x,ˆx) = if x ˆx Consider any joint distribution that satisfies the distortion constraint D. Since D = Pr(X ˆX), we have by Fano s inequality and hence H(X ˆX) H(D)+Dlog(m ), (9) I(X; ˆX) = H(X) H(X ˆX) (0) logm H(D) Dlog(m ). () 5

6 We can achieve this lower bound by choosing p(ˆx) to be the uniform distribution, and the conditional distribution of p(x ˆx) to be p(ˆx x) { = D if ˆx = x = D/(m ) if ˆx x. () It is easy to verify that this gives the right distribution on X and satisfies the bound with equality for D < m. Hence { = logm H(D) Dlog(m ) if 0 D R(D) m 0 if D >. (3) m 5. Erasure distortion Consider X Bernoulli( ), and let the distortion measure be given by the matrix [ ] 0 d(x,ˆx) =. (4) 0 Calculate the rate distortion function for this source. Can you suggest a simple scheme to achieve any value of the rate distortion function for this source? Solution: Erasure distortion The infinite distortion constrains p(0,) = p(,0) = 0. Hence by symmetry the joint distribution of (X, ˆX) is of the form shown in Figure. 0 a 0 a e a a Figure : Joint distribution for erasure rate distortion of a binary source. 6

7 For this joint distribution, it is easy to calculate the distortion D = a and that I(X; ˆX) = H(X) H(X ˆX) = a. Hence we have R(D) = D for 0 D. For D >, R(D) = 0. It is very see how we could achieve this rate distortion function. If D is rational, say k/n, then we send only the first n k of any block of n bits. We reproduce these bits exactly and reproduce the remaining bits as erasures. Hence we can send information at rate D and achieve a distortion D. If D is irrational, we can get arbitrarily close to D by using longer and longer block lengths. 6. Rate distortion. A memoryless source U is uniformly distributed on {0,...,r }. The following distortion function is given by d(u,v) = 0, u = v,, u = v ± mod r,, otherwise. Show that the rate distortion function is { logr D h (D), D 3 R(D) =, logr log3, D >. 3 Solution: Rate distortion From the symmetry of the problem, we can assume the conditional distribution of p(v u) as p u = v, p(v u) = p/ u = v ± mod r, 0 otherwise Then, E(d(U,V)) = p. Therefore, the rate distortion function is Now, we know that R(D) = min p D I(U;V). I(U;V) =H(V) H(V U) =logr H( p,p/,p/), 7

8 since U is uniform, and due to symmetry, V is also uniform. We know that H( p,p,p) log3, and this is achieved when p = /3. Therefore, R(D) = logr log3, if D > /3. Now, let s consider the case when D /3. Denote f(p) = H( p,p/,p/) = ( p)log( p) p/logp/ = ( p)log( p) plogp+p, We know that f(p) is a concave function. By differentiating with respect to p, df(p) dp = log( p)+ logp + = log p p + and setting f(p) = 0, f(p) becomes maximum when p = /3. Therefore, if D /3, f(p) is an increasing function of p. Thus,. R(D) = log3 D h (D), if D /3. 7. Adding a column to the distortion matrix. Let R(D) be the rate distortion function for an i.i.d. process with probability mass function p(x) and distortion function d(x,ˆx), x X, ˆx ˆX. Now suppose that we add a new reproduction symbol ˆx 0 to ˆX with associated distortion d(x,ˆx 0 ), x X. Can this increase R(D)? Explain. Solution: Adding a column Let the new rate distortion function be denoted as R(D), and note that we can still achieve R(D) by restricting the support of p(x,ˆx), i.e., by simply ignoring the new symbol. Thus, R(D) R(D). Finally note the duality to the problem in which we added a row to the channel transition matrix to have no smaller capacity (Problem 7.). 8. Simplification. Suppose X = {,,3,4}, ˆX = {,,3,4}, p(i) = 4, i =,,3,4, and X,X,... are i.i.d. p(x). The distortion matrix 8

9 d(x,ˆx) is given by (a) Find R(0), the rate necessary to describe the process with zero distortion. (b) Find the rate distortion function R(D). (Hint: The distortion measure allows to simplify the problem into one you have already seen.) (c) Suppose we have a nonuniform distribution p(i) = p i, i =,,3,4. What is R(D)? Solution: Simplification (a) We can achieve 0 distortion if we output ˆX = if X = or, and ˆX = 3 if X = 3 or 4. Thus if we set Y = if X = or, and Y = if X = 3 or 4, we can recover Y exactly if the rate is greater than H(Y) = bit. It is also not hard to see that any 0 distortion code would be able to recover Y exactly, and thus R(0) =. (b) If we define Y as in the previous part, and Ŷ similarly from ˆX, we can see that the distortion between X and ˆX is equal to the Hamming distortion between Y and Ŷ. Therefore if the rate is greater than the Hamming rate distortion function R(D) for Y, we can recover X to distortion D. Thus R(D) = H(D). (c) If the distribution of X is not uniform, the same arguments hold and Y has a distribution (p +p,p 3 +p 4 ), and the rate distortion function is R(D) = H(p +p ) H(D), 9. Rate distortion for two independent sources. Can one simultaneously compress two independent sources better than compressing the sources individually? The following problem addresses this question. Let the pair {(X i,y i )} be iid p(x,y). The distortion measure for X is d(x,ˆx) and its rate distortion function is R X (D). Similarly, the 9

10 distortion measure for Y is d(y,ŷ) and its rate distortion function is R Y (D). Suppose we now wish to describe the process {(X i,y i )} subject to distortionconstraintslim n Ed(X n, ˆX n ) D andlim n Ed(Y n,ŷ n ) D. Our rate distortion theorem can be shown to naturally extend to this setting and imply that the minimum rate required to achieve these distortion is given by R X,Y (D,D ) = min p(ˆx,ŷ x,y):ed(x, ˆX) D,Ed(Y,Ŷ) D I(X,Y; ˆX,Ŷ) Now, suppose the {X i } process and the {Y i } process are independent of each other. (a) Show (b) Does equality hold? Now answer the question. R X,Y (D,D ) R X (D )+R Y (D ). Solution: Rate distortion for two independent sources. (a) Given that X and Y are independent, we have Then I(X,Y; p(x,y,ˆx,ŷ) = p(x)p(y)p(ˆx,ŷ x,y) (5) ˆX,Ŷ) = H(X,Y) H(X,Y ˆX,Ŷ) (6) = H(X)+H(Y) H(X ˆX,Ŷ) H(Y X, ˆX,Ŷ) H(X)+H(Y) H(X ˆX) H(Y Ŷ) (8) = I(X; ˆX)+I(Y;Ŷ) (9) where the inequality follows from the fact that conditioning reduces entropy. Therefore R X,Y (D,D ) = min I(X,Y; ˆX,Ŷ) (30) p(ˆx,ŷ x,y):ed(x, ˆX) D,Ed(Y,Ŷ) D min p(ˆx,ŷ x,y):ed(x, ˆX) D,Ed(Y,Ŷ) D (I(X; ˆX)+I(Y;Ŷ) ) (3) = min I(X; ˆX)+ min I(Y;Ŷ) (3) p(ˆx x):ed(x, ˆX) D p(ŷ y):ed(y,ŷ) D = R X (D )+R Y (D ) (33) 0

11 (b) If then I(X,Y; p(x, y, ˆx, ŷ) = p(x)p(y)p(ˆx x)p(ŷ y), (34) ˆX,Ŷ) = H(X,Y) H(X,Y ˆX,Ŷ) (35) = H(X)+H(Y) H(X ˆX,Ŷ) H(Y X, ˆX,Ŷ) = H(X)+H(Y) H(X ˆX) H(Y Ŷ) (37) = I(X; ˆX)+I(Y;Ŷ) (38) Let p(x, ˆx) be a distribution that achieves the rate distortion R X (D ) at distortion D and let p(y,ŷ) be a distribution that achieves the ratedistortionr Y (D ) atdistortion D. Then forthe product distribution p(x,y,ˆx,ŷ) = p(x,ˆx)p(y,ŷ), where the componentdistributionsachieverates(d,r X (D ))and(d,r X (D )), the mutual information corresponding to the product distribution is R X (D )+R Y (D ). Thus R X,Y (D,D ) = min I(X,Y; ˆX,Ŷ) = R X(D )+R Y (D ) p(ˆx,ŷ x,y):ed(x, ˆX) D,Ed(Y,Ŷ) D (39) Thus by using the product distribution, we can achieve the sum of the rates. Therefore the total rate at which we encode two independent sources together with distortions D and D is the same as if we encoded each of them separately. 0. One bit quantization of a single Gaussian random variable. Let X Norm(0,σ ) and let the distortion measure be squared error. Here we do not allow block descriptions. Show that the optimum reproduction points for bit quantization are ± π σ, and that the expected distortion for bit quantization is π π σ. Compare this with the distortion rate bound D = σ R for R =. Solution: One bit quantization of a single Gaussian random variable With one bit quantization, the obvious reconstruction regions are the positive and negative real axes. The reconstruction point is the centroid

12 of each region. For example, for the positive real line, the centroid a is given as a = x 0 πσ = σ 0 π e y dy = σ π, e x σ dx using the substitution y = x /σ. The expected distortion for one bit quantization is D = 0 ( x+σ π ( + x σ 0 π ( x +σ ) π πσ ( ) xσ π = 0 = σ + π σ 4 π σ π = σ π π.3634σ, ) x e σ dx πσ ) x e σ dx πσ e x σ dx x e σ dx πσ which is much larger than the distortion rate bound D() = σ /4.. Side information. Amemorylesssourcegeneratesi.i.d.pairsofrandomvariables(U i,s i ), i =,,... on finite alphabets, according to p(u n,s n ) = n p(u i,s i ). i=

13 We are interested in describing U, when S, called side information, is known both at the encoder and at the decoder. Prove that the rate distortion function is given by R U S (D) = min I(U;V S). p(v u,s):e[d(u,v)] D Compare R U S (D) with the ordinary rate distortion function R U (D) without any side information. What can you say about the influence of the correlation between U and S on R U S (D)? Solution: Side information Since both encoder and decoder have the common side information S, we can be helped from the correlation of our source U and the side information S. The idea for the achievability is, we use the different rate distortion code for the source symbols that have different corresponding side information. Therefore, for the symbols that has side information S = s will have the rate distortion code at rate R U S=s (D) = min I(U;V S = s) p(v u,s=s):e(d(u,v)) D and expected distortion D. Now, the total rate will be S R = P S n(s)r U S=s (D), s= where P S n(s) is the empirical distribution of S n. As n becomes large enough, we know that P S n(s) will be very close to the true distribution of S, p(s), or with high probability, S n will be in the strong typical set T P (δ), where { T P (δ) = s n : max P s n(s) p(s) < δ }. s S S Therefore, if n sufficiently large, and δ sufficiently small, we will achieve the rate R = p(s)i(u;v S = s) = I(U;V S) s S 3

14 It is clear that the overall rate distortion code has the expected distortion D. For the converse, we have nr H(V n S n ) =I(U n ;V n S n ) =H(U n S n ) H(U n V n,s n ) ( = H(Ui S i ) H(U i U i,v n,s n ) ) = i= (H(U i S i ) H(U i V i )) i= I(U i ;V i S i ) i= R U S (Ed(U i,v i )) i= R U S (Ed(U n,v n )) R U S (D) where each step has the identical reason as the original converse. Now, to compare R U S (D) and R U (D), consider following: R U S (D) = min p(v u,s):e(d(u,v)) D min p(v u):e(d(u,v)) D min p(v u):e(d(u,v)) D =R U (D). I(U;V S) I(U;V S) (40) I(U;V) (4) Here, the inequality in (40) is from the fact that the minimization is held on the smaller set. The inequality in (4) is from the fact that S U V forms a Markov chain with the joint distribution that we get from the minimization in (40). Also, we have, I(U,S;V) =I(U;V)+I(S;V U) =I(S;V)+I(U;V S). 4

15 Since I(S;V U) = 0 and I(S;V) 0, we have the inequality in (4). Therefore, we can only gain from the the side information. Intuitively, this also makes sense, because we can always just ignore the side information and get the original rate distortion code, but the rate may be worse. The two rates will be equal if and only if U and S are independent, which achieves the equality in both (40) and (4). 5

Lecture 22: Final Review

Lecture 22: Final Review Lecture 22: Final Review Nuts and bolts Fundamental questions and limits Tools Practical algorithms Future topics Dr Yao Xie, ECE587, Information Theory, Duke University Basics Dr Yao Xie, ECE587, Information

More information

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University Chapter 4 Data Transmission and Channel Capacity Po-Ning Chen, Professor Department of Communications Engineering National Chiao Tung University Hsin Chu, Taiwan 30050, R.O.C. Principle of Data Transmission

More information

Lecture 3: Channel Capacity

Lecture 3: Channel Capacity Lecture 3: Channel Capacity 1 Definitions Channel capacity is a measure of maximum information per channel usage one can get through a channel. This one of the fundamental concepts in information theory.

More information

National University of Singapore Department of Electrical & Computer Engineering. Examination for

National University of Singapore Department of Electrical & Computer Engineering. Examination for National University of Singapore Department of Electrical & Computer Engineering Examination for EE5139R Information Theory for Communication Systems (Semester I, 2014/15) November/December 2014 Time Allowed:

More information

EE5319R: Problem Set 3 Assigned: 24/08/16, Due: 31/08/16

EE5319R: Problem Set 3 Assigned: 24/08/16, Due: 31/08/16 EE539R: Problem Set 3 Assigned: 24/08/6, Due: 3/08/6. Cover and Thomas: Problem 2.30 (Maimum Entropy): Solution: We are required to maimize H(P X ) over all distributions P X on the non-negative integers

More information

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye Chapter 2: Entropy and Mutual Information Chapter 2 outline Definitions Entropy Joint entropy, conditional entropy Relative entropy, mutual information Chain rules Jensen s inequality Log-sum inequality

More information

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

Lecture 6 I. CHANNEL CODING. X n (m) P Y X 6- Introduction to Information Theory Lecture 6 Lecturer: Haim Permuter Scribe: Yoav Eisenberg and Yakov Miron I. CHANNEL CODING We consider the following channel coding problem: m = {,2,..,2 nr} Encoder

More information

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions EE/Stat 376B Handout #5 Network Information Theory October, 14, 014 1. Problem.4 parts (b) and (c). Homework Set # Solutions (b) Consider h(x + Y ) h(x + Y Y ) = h(x Y ) = h(x). (c) Let ay = Y 1 + Y, where

More information

ECE Information theory Final (Fall 2008)

ECE Information theory Final (Fall 2008) ECE 776 - Information theory Final (Fall 2008) Q.1. (1 point) Consider the following bursty transmission scheme for a Gaussian channel with noise power N and average power constraint P (i.e., 1/n X n i=1

More information

UCSD ECE 255C Handout #14 Prof. Young-Han Kim Thursday, March 9, Solutions to Homework Set #5

UCSD ECE 255C Handout #14 Prof. Young-Han Kim Thursday, March 9, Solutions to Homework Set #5 UCSD ECE 255C Handout #14 Prof. Young-Han Kim Thursday, March 9, 2017 Solutions to Homework Set #5 3.18 Bounds on the quadratic rate distortion function. Recall that R(D) = inf F(ˆx x):e(x ˆX)2 DI(X; ˆX).

More information

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018 Please submit the solutions on Gradescope. EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018 1. Optimal codeword lengths. Although the codeword lengths of an optimal variable length code

More information

Lecture 5: Channel Capacity. Copyright G. Caire (Sample Lectures) 122

Lecture 5: Channel Capacity. Copyright G. Caire (Sample Lectures) 122 Lecture 5: Channel Capacity Copyright G. Caire (Sample Lectures) 122 M Definitions and Problem Setup 2 X n Y n Encoder p(y x) Decoder ˆM Message Channel Estimate Definition 11. Discrete Memoryless Channel

More information

Electrical and Information Technology. Information Theory. Problems and Solutions. Contents. Problems... 1 Solutions...7

Electrical and Information Technology. Information Theory. Problems and Solutions. Contents. Problems... 1 Solutions...7 Electrical and Information Technology Information Theory Problems and Solutions Contents Problems.......... Solutions...........7 Problems 3. In Problem?? the binomial coefficent was estimated with Stirling

More information

Chapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye

Chapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye Chapter 8: Differential entropy Chapter 8 outline Motivation Definitions Relation to discrete entropy Joint and conditional differential entropy Relative entropy and mutual information Properties AEP for

More information

Solutions to Set #2 Data Compression, Huffman code and AEP

Solutions to Set #2 Data Compression, Huffman code and AEP Solutions to Set #2 Data Compression, Huffman code and AEP. Huffman coding. Consider the random variable ( ) x x X = 2 x 3 x 4 x 5 x 6 x 7 0.50 0.26 0. 0.04 0.04 0.03 0.02 (a) Find a binary Huffman code

More information

(Classical) Information Theory III: Noisy channel coding

(Classical) Information Theory III: Noisy channel coding (Classical) Information Theory III: Noisy channel coding Sibasish Ghosh The Institute of Mathematical Sciences CIT Campus, Taramani, Chennai 600 113, India. p. 1 Abstract What is the best possible way

More information

ECE Information theory Final

ECE Information theory Final ECE 776 - Information theory Final Q1 (1 point) We would like to compress a Gaussian source with zero mean and variance 1 We consider two strategies In the first, we quantize with a step size so that the

More information

SOURCE CODING WITH SIDE INFORMATION AT THE DECODER (WYNER-ZIV CODING) FEB 13, 2003

SOURCE CODING WITH SIDE INFORMATION AT THE DECODER (WYNER-ZIV CODING) FEB 13, 2003 SOURCE CODING WITH SIDE INFORMATION AT THE DECODER (WYNER-ZIV CODING) FEB 13, 2003 SLEPIAN-WOLF RESULT { X i} RATE R x ENCODER 1 DECODER X i V i {, } { V i} ENCODER 0 RATE R v Problem: Determine R, the

More information

ECE 4400:693 - Information Theory

ECE 4400:693 - Information Theory ECE 4400:693 - Information Theory Dr. Nghi Tran Lecture 8: Differential Entropy Dr. Nghi Tran (ECE-University of Akron) ECE 4400:693 Lecture 1 / 43 Outline 1 Review: Entropy of discrete RVs 2 Differential

More information

Solutions to Homework Set #3 Channel and Source coding

Solutions to Homework Set #3 Channel and Source coding Solutions to Homework Set #3 Channel and Source coding. Rates (a) Channels coding Rate: Assuming you are sending 4 different messages using usages of a channel. What is the rate (in bits per channel use)

More information

Network coding for multicast relation to compression and generalization of Slepian-Wolf

Network coding for multicast relation to compression and generalization of Slepian-Wolf Network coding for multicast relation to compression and generalization of Slepian-Wolf 1 Overview Review of Slepian-Wolf Distributed network compression Error exponents Source-channel separation issues

More information

INFORMATION THEORY AND STATISTICS

INFORMATION THEORY AND STATISTICS CHAPTER INFORMATION THEORY AND STATISTICS We now explore the relationship between information theory and statistics. We begin by describing the method of types, which is a powerful technique in large deviation

More information

Capacity of a channel Shannon s second theorem. Information Theory 1/33

Capacity of a channel Shannon s second theorem. Information Theory 1/33 Capacity of a channel Shannon s second theorem Information Theory 1/33 Outline 1. Memoryless channels, examples ; 2. Capacity ; 3. Symmetric channels ; 4. Channel Coding ; 5. Shannon s second theorem,

More information

18.2 Continuous Alphabet (discrete-time, memoryless) Channel

18.2 Continuous Alphabet (discrete-time, memoryless) Channel 0-704: Information Processing and Learning Spring 0 Lecture 8: Gaussian channel, Parallel channels and Rate-distortion theory Lecturer: Aarti Singh Scribe: Danai Koutra Disclaimer: These notes have not

More information

Solutions to Homework Set #2 Broadcast channel, degraded message set, Csiszar Sum Equality

Solutions to Homework Set #2 Broadcast channel, degraded message set, Csiszar Sum Equality 1st Semester 2010/11 Solutions to Homework Set #2 Broadcast channel, degraded message set, Csiszar Sum Equality 1. Convexity of capacity region of broadcast channel. Let C R 2 be the capacity region of

More information

Lecture 4 Noisy Channel Coding

Lecture 4 Noisy Channel Coding Lecture 4 Noisy Channel Coding I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw October 9, 2015 1 / 56 I-Hsiang Wang IT Lecture 4 The Channel Coding Problem

More information

Quiz 2 Date: Monday, November 21, 2016

Quiz 2 Date: Monday, November 21, 2016 10-704 Information Processing and Learning Fall 2016 Quiz 2 Date: Monday, November 21, 2016 Name: Andrew ID: Department: Guidelines: 1. PLEASE DO NOT TURN THIS PAGE UNTIL INSTRUCTED. 2. Write your name,

More information

EE5585 Data Compression May 2, Lecture 27

EE5585 Data Compression May 2, Lecture 27 EE5585 Data Compression May 2, 2013 Lecture 27 Instructor: Arya Mazumdar Scribe: Fangying Zhang Distributed Data Compression/Source Coding In the previous class we used a H-W table as a simple example,

More information

EE376A: Homeworks #4 Solutions Due on Thursday, February 22, 2018 Please submit on Gradescope. Start every question on a new page.

EE376A: Homeworks #4 Solutions Due on Thursday, February 22, 2018 Please submit on Gradescope. Start every question on a new page. EE376A: Homeworks #4 Solutions Due on Thursday, February 22, 28 Please submit on Gradescope. Start every question on a new page.. Maximum Differential Entropy (a) Show that among all distributions supported

More information

EE376A - Information Theory Final, Monday March 14th 2016 Solutions. Please start answering each question on a new page of the answer booklet.

EE376A - Information Theory Final, Monday March 14th 2016 Solutions. Please start answering each question on a new page of the answer booklet. EE376A - Information Theory Final, Monday March 14th 216 Solutions Instructions: You have three hours, 3.3PM - 6.3PM The exam has 4 questions, totaling 12 points. Please start answering each question on

More information

Exercise 1. = P(y a 1)P(a 1 )

Exercise 1. = P(y a 1)P(a 1 ) Chapter 7 Channel Capacity Exercise 1 A source produces independent, equally probable symbols from an alphabet {a 1, a 2 } at a rate of one symbol every 3 seconds. These symbols are transmitted over a

More information

X 1 : X Table 1: Y = X X 2

X 1 : X Table 1: Y = X X 2 ECE 534: Elements of Information Theory, Fall 200 Homework 3 Solutions (ALL DUE to Kenneth S. Palacio Baus) December, 200. Problem 5.20. Multiple access (a) Find the capacity region for the multiple-access

More information

EE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15

EE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15 EE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15 1. Cascade of Binary Symmetric Channels The conditional probability distribution py x for each of the BSCs may be expressed by the transition probability

More information

CHAPTER 3. P (B j A i ) P (B j ) =log 2. j=1

CHAPTER 3. P (B j A i ) P (B j ) =log 2. j=1 CHAPTER 3 Problem 3. : Also : Hence : I(B j ; A i ) = log P (B j A i ) P (B j ) 4 P (B j )= P (B j,a i )= i= 3 P (A i )= P (B j,a i )= j= =log P (B j,a i ) P (B j )P (A i ).3, j=.7, j=.4, j=3.3, i=.7,

More information

ELEC546 Review of Information Theory

ELEC546 Review of Information Theory ELEC546 Review of Information Theory Vincent Lau 1/1/004 1 Review of Information Theory Entropy: Measure of uncertainty of a random variable X. The entropy of X, H(X), is given by: If X is a discrete random

More information

Principles of Communications

Principles of Communications Principles of Communications Weiyao Lin Shanghai Jiao Tong University Chapter 10: Information Theory Textbook: Chapter 12 Communication Systems Engineering: Ch 6.1, Ch 9.1~ 9. 92 2009/2010 Meixia Tao @

More information

Homework Set #2 Data Compression, Huffman code and AEP

Homework Set #2 Data Compression, Huffman code and AEP Homework Set #2 Data Compression, Huffman code and AEP 1. Huffman coding. Consider the random variable ( x1 x X = 2 x 3 x 4 x 5 x 6 x 7 0.50 0.26 0.11 0.04 0.04 0.03 0.02 (a Find a binary Huffman code

More information

UCSD ECE 255C Handout #12 Prof. Young-Han Kim Tuesday, February 28, Solutions to Take-Home Midterm (Prepared by Pinar Sen)

UCSD ECE 255C Handout #12 Prof. Young-Han Kim Tuesday, February 28, Solutions to Take-Home Midterm (Prepared by Pinar Sen) UCSD ECE 255C Handout #12 Prof. Young-Han Kim Tuesday, February 28, 2017 Solutions to Take-Home Midterm (Prepared by Pinar Sen) 1. (30 points) Erasure broadcast channel. Let p(y 1,y 2 x) be a discrete

More information

Noisy-Channel Coding

Noisy-Channel Coding Copyright Cambridge University Press 2003. On-screen viewing permitted. Printing not permitted. http://www.cambridge.org/05264298 Part II Noisy-Channel Coding Copyright Cambridge University Press 2003.

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science Transmission of Information Spring 2006

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science Transmission of Information Spring 2006 MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.44 Transmission of Information Spring 2006 Homework 2 Solution name username April 4, 2006 Reading: Chapter

More information

Lecture 8: Channel Capacity, Continuous Random Variables

Lecture 8: Channel Capacity, Continuous Random Variables EE376A/STATS376A Information Theory Lecture 8-02/0/208 Lecture 8: Channel Capacity, Continuous Random Variables Lecturer: Tsachy Weissman Scribe: Augustine Chemparathy, Adithya Ganesh, Philip Hwang Channel

More information

Ch. 8 Math Preliminaries for Lossy Coding. 8.4 Info Theory Revisited

Ch. 8 Math Preliminaries for Lossy Coding. 8.4 Info Theory Revisited Ch. 8 Math Preliminaries for Lossy Coding 8.4 Info Theory Revisited 1 Info Theory Goals for Lossy Coding Again just as for the lossless case Info Theory provides: Basis for Algorithms & Bounds on Performance

More information

Lecture 15: Conditional and Joint Typicaility

Lecture 15: Conditional and Joint Typicaility EE376A Information Theory Lecture 1-02/26/2015 Lecture 15: Conditional and Joint Typicaility Lecturer: Kartik Venkat Scribe: Max Zimet, Brian Wai, Sepehr Nezami 1 Notation We always write a sequence of

More information

Lecture 2. Capacity of the Gaussian channel

Lecture 2. Capacity of the Gaussian channel Spring, 207 5237S, Wireless Communications II 2. Lecture 2 Capacity of the Gaussian channel Review on basic concepts in inf. theory ( Cover&Thomas: Elements of Inf. Theory, Tse&Viswanath: Appendix B) AWGN

More information

LECTURE 3. Last time:

LECTURE 3. Last time: LECTURE 3 Last time: Mutual Information. Convexity and concavity Jensen s inequality Information Inequality Data processing theorem Fano s Inequality Lecture outline Stochastic processes, Entropy rate

More information

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16 EE539R: Problem Set 4 Assigned: 3/08/6, Due: 07/09/6. Cover and Thomas: Problem 3.5 Sets defined by probabilities: Define the set C n (t = {x n : P X n(x n 2 nt } (a We have = P X n(x n P X n(x n 2 nt

More information

4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information

4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information 4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information Ramji Venkataramanan Signal Processing and Communications Lab Department of Engineering ramji.v@eng.cam.ac.uk

More information

Universal Incremental Slepian-Wolf Coding

Universal Incremental Slepian-Wolf Coding Proceedings of the 43rd annual Allerton Conference, Monticello, IL, September 2004 Universal Incremental Slepian-Wolf Coding Stark C. Draper University of California, Berkeley Berkeley, CA, 94720 USA sdraper@eecs.berkeley.edu

More information

An Extended Fano s Inequality for the Finite Blocklength Coding

An Extended Fano s Inequality for the Finite Blocklength Coding An Extended Fano s Inequality for the Finite Bloclength Coding Yunquan Dong, Pingyi Fan {dongyq8@mails,fpy@mail}.tsinghua.edu.cn Department of Electronic Engineering, Tsinghua University, Beijing, P.R.

More information

Midterm Exam Information Theory Fall Midterm Exam. Time: 09:10 12:10 11/23, 2016

Midterm Exam Information Theory Fall Midterm Exam. Time: 09:10 12:10 11/23, 2016 Midterm Exam Time: 09:10 12:10 11/23, 2016 Name: Student ID: Policy: (Read before You Start to Work) The exam is closed book. However, you are allowed to bring TWO A4-size cheat sheet (single-sheet, two-sided).

More information

Exercises with solutions (Set B)

Exercises with solutions (Set B) Exercises with solutions (Set B) 3. A fair coin is tossed an infinite number of times. Let Y n be a random variable, with n Z, that describes the outcome of the n-th coin toss. If the outcome of the n-th

More information

(each row defines a probability distribution). Given n-strings x X n, y Y n we can use the absence of memory in the channel to compute

(each row defines a probability distribution). Given n-strings x X n, y Y n we can use the absence of memory in the channel to compute ENEE 739C: Advanced Topics in Signal Processing: Coding Theory Instructor: Alexander Barg Lecture 6 (draft; 9/6/03. Error exponents for Discrete Memoryless Channels http://www.enee.umd.edu/ abarg/enee739c/course.html

More information

Lecture 11: Continuous-valued signals and differential entropy

Lecture 11: Continuous-valued signals and differential entropy Lecture 11: Continuous-valued signals and differential entropy Biology 429 Carl Bergstrom September 20, 2008 Sources: Parts of today s lecture follow Chapter 8 from Cover and Thomas (2007). Some components

More information

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 ECE598: Information-theoretic methods in high-dimensional statistics Spring 06 Lecture : Mutual Information Method Lecturer: Yihong Wu Scribe: Jaeho Lee, Mar, 06 Ed. Mar 9 Quick review: Assouad s lemma

More information

Note that the new channel is noisier than the original two : and H(A I +A2-2A1A2) > H(A2) (why?). min(c,, C2 ) = min(1 - H(a t ), 1 - H(A 2 )).

Note that the new channel is noisier than the original two : and H(A I +A2-2A1A2) > H(A2) (why?). min(c,, C2 ) = min(1 - H(a t ), 1 - H(A 2 )). l I ~-16 / (a) (5 points) What is the capacity Cr of the channel X -> Y? What is C of the channel Y - Z? (b) (5 points) What is the capacity C 3 of the cascaded channel X -3 Z? (c) (5 points) A ow let.

More information

Lecture 2: August 31

Lecture 2: August 31 0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 2: August 3 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy

More information

Shannon s Noisy-Channel Coding Theorem

Shannon s Noisy-Channel Coding Theorem Shannon s Noisy-Channel Coding Theorem Lucas Slot Sebastian Zur February 2015 Abstract In information theory, Shannon s Noisy-Channel Coding Theorem states that it is possible to communicate over a noisy

More information

Lecture 10: Broadcast Channel and Superposition Coding

Lecture 10: Broadcast Channel and Superposition Coding Lecture 10: Broadcast Channel and Superposition Coding Scribed by: Zhe Yao 1 Broadcast channel M 0M 1M P{y 1 y x} M M 01 1 M M 0 The capacity of the broadcast channel depends only on the marginal conditional

More information

Entropies & Information Theory

Entropies & Information Theory Entropies & Information Theory LECTURE I Nilanjana Datta University of Cambridge,U.K. See lecture notes on: http://www.qi.damtp.cam.ac.uk/node/223 Quantum Information Theory Born out of Classical Information

More information

Lecture 8: Channel and source-channel coding theorems; BEC & linear codes. 1 Intuitive justification for upper bound on channel capacity

Lecture 8: Channel and source-channel coding theorems; BEC & linear codes. 1 Intuitive justification for upper bound on channel capacity 5-859: Information Theory and Applications in TCS CMU: Spring 23 Lecture 8: Channel and source-channel coding theorems; BEC & linear codes February 7, 23 Lecturer: Venkatesan Guruswami Scribe: Dan Stahlke

More information

Remote Source Coding with Two-Sided Information

Remote Source Coding with Two-Sided Information Remote Source Coding with Two-Sided Information Basak Guler Ebrahim MolavianJazi Aylin Yener Wireless Communications and Networking Laboratory Department of Electrical Engineering The Pennsylvania State

More information

Noisy channel communication

Noisy channel communication Information Theory http://www.inf.ed.ac.uk/teaching/courses/it/ Week 6 Communication channels and Information Some notes on the noisy channel setup: Iain Murray, 2012 School of Informatics, University

More information

LECTURE 13. Last time: Lecture outline

LECTURE 13. Last time: Lecture outline LECTURE 13 Last time: Strong coding theorem Revisiting channel and codes Bound on probability of error Error exponent Lecture outline Fano s Lemma revisited Fano s inequality for codewords Converse to

More information

The Gallager Converse

The Gallager Converse The Gallager Converse Abbas El Gamal Director, Information Systems Laboratory Department of Electrical Engineering Stanford University Gallager s 75th Birthday 1 Information Theoretic Limits Establishing

More information

Lecture 14 February 28

Lecture 14 February 28 EE/Stats 376A: Information Theory Winter 07 Lecture 4 February 8 Lecturer: David Tse Scribe: Sagnik M, Vivek B 4 Outline Gaussian channel and capacity Information measures for continuous random variables

More information

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1 Kraft s inequality An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if N 2 l i 1 Proof: Suppose that we have a tree code. Let l max = max{l 1,...,

More information

EE 4TM4: Digital Communications II. Channel Capacity

EE 4TM4: Digital Communications II. Channel Capacity EE 4TM4: Digital Communications II 1 Channel Capacity I. CHANNEL CODING THEOREM Definition 1: A rater is said to be achievable if there exists a sequence of(2 nr,n) codes such thatlim n P (n) e (C) = 0.

More information

LECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem

LECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem LECTURE 15 Last time: Feedback channel: setting up the problem Perfect feedback Feedback capacity Data compression Lecture outline Joint source and channel coding theorem Converse Robustness Brain teaser

More information

Chapter 9 Fundamental Limits in Information Theory

Chapter 9 Fundamental Limits in Information Theory Chapter 9 Fundamental Limits in Information Theory Information Theory is the fundamental theory behind information manipulation, including data compression and data transmission. 9.1 Introduction o For

More information

MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK. SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A

MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK. SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK DEPARTMENT: ECE SEMESTER: IV SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A 1. What is binary symmetric channel (AUC DEC

More information

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information. L65 Dept. of Linguistics, Indiana University Fall 205 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission rate

More information

Dept. of Linguistics, Indiana University Fall 2015

Dept. of Linguistics, Indiana University Fall 2015 L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 28 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission

More information

On Optimum Conventional Quantization for Source Coding with Side Information at the Decoder

On Optimum Conventional Quantization for Source Coding with Side Information at the Decoder On Optimum Conventional Quantization for Source Coding with Side Information at the Decoder by Lin Zheng A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the

More information

Lecture 17: Differential Entropy

Lecture 17: Differential Entropy Lecture 17: Differential Entropy Differential entropy AEP for differential entropy Quantization Maximum differential entropy Estimation counterpart of Fano s inequality Dr. Yao Xie, ECE587, Information

More information

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions Engineering Tripos Part IIA THIRD YEAR 3F: Signals and Systems INFORMATION THEORY Examples Paper Solutions. Let the joint probability mass function of two binary random variables X and Y be given in the

More information

The Communication Complexity of Correlation. Prahladh Harsha Rahul Jain David McAllester Jaikumar Radhakrishnan

The Communication Complexity of Correlation. Prahladh Harsha Rahul Jain David McAllester Jaikumar Radhakrishnan The Communication Complexity of Correlation Prahladh Harsha Rahul Jain David McAllester Jaikumar Radhakrishnan Transmitting Correlated Variables (X, Y) pair of correlated random variables Transmitting

More information

Lecture 3. Mathematical methods in communication I. REMINDER. A. Convex Set. A set R is a convex set iff, x 1,x 2 R, θ, 0 θ 1, θx 1 + θx 2 R, (1)

Lecture 3. Mathematical methods in communication I. REMINDER. A. Convex Set. A set R is a convex set iff, x 1,x 2 R, θ, 0 θ 1, θx 1 + θx 2 R, (1) 3- Mathematical methods in communication Lecture 3 Lecturer: Haim Permuter Scribe: Yuval Carmel, Dima Khaykin, Ziv Goldfeld I. REMINDER A. Convex Set A set R is a convex set iff, x,x 2 R, θ, θ, θx + θx

More information

MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006)

MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006) MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK SATELLITE COMMUNICATION DEPT./SEM.:ECE/VIII UNIT V PART-A 1. What is binary symmetric channel (AUC DEC 2006) 2. Define information rate? (AUC DEC 2007)

More information

The Method of Types and Its Application to Information Hiding

The Method of Types and Its Application to Information Hiding The Method of Types and Its Application to Information Hiding Pierre Moulin University of Illinois at Urbana-Champaign www.ifp.uiuc.edu/ moulin/talks/eusipco05-slides.pdf EUSIPCO Antalya, September 7,

More information

Coding for Discrete Source

Coding for Discrete Source EGR 544 Communication Theory 3. Coding for Discrete Sources Z. Aliyazicioglu Electrical and Computer Engineering Department Cal Poly Pomona Coding for Discrete Source Coding Represent source data effectively

More information

CS6304 / Analog and Digital Communication UNIT IV - SOURCE AND ERROR CONTROL CODING PART A 1. What is the use of error control coding? The main use of error control coding is to reduce the overall probability

More information

Notes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel

Notes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel Introduction to Coding Theory CMU: Spring 2010 Notes 3: Stochastic channels and noisy coding theorem bound January 2010 Lecturer: Venkatesan Guruswami Scribe: Venkatesan Guruswami We now turn to the basic

More information

Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels

Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels Nan Liu and Andrea Goldsmith Department of Electrical Engineering Stanford University, Stanford CA 94305 Email:

More information

Common Information. Abbas El Gamal. Stanford University. Viterbi Lecture, USC, April 2014

Common Information. Abbas El Gamal. Stanford University. Viterbi Lecture, USC, April 2014 Common Information Abbas El Gamal Stanford University Viterbi Lecture, USC, April 2014 Andrew Viterbi s Fabulous Formula, IEEE Spectrum, 2010 El Gamal (Stanford University) Disclaimer Viterbi Lecture 2

More information

Lecture 5 Channel Coding over Continuous Channels

Lecture 5 Channel Coding over Continuous Channels Lecture 5 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 14, 2014 1 / 34 I-Hsiang Wang NIT Lecture 5 From

More information

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria Source Coding Master Universitario en Ingeniería de Telecomunicación I. Santamaría Universidad de Cantabria Contents Introduction Asymptotic Equipartition Property Optimal Codes (Huffman Coding) Universal

More information

Example: Letter Frequencies

Example: Letter Frequencies Example: Letter Frequencies i a i p i 1 a 0.0575 2 b 0.0128 3 c 0.0263 4 d 0.0285 5 e 0.0913 6 f 0.0173 7 g 0.0133 8 h 0.0313 9 i 0.0599 10 j 0.0006 11 k 0.0084 12 l 0.0335 13 m 0.0235 14 n 0.0596 15 o

More information

Introduction to Information Theory. B. Škorić, Physical Aspects of Digital Security, Chapter 2

Introduction to Information Theory. B. Škorić, Physical Aspects of Digital Security, Chapter 2 Introduction to Information Theory B. Škorić, Physical Aspects of Digital Security, Chapter 2 1 Information theory What is it? - formal way of counting information bits Why do we need it? - often used

More information

Lecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157

Lecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157 Lecture 6: Gaussian Channels Copyright G. Caire (Sample Lectures) 157 Differential entropy (1) Definition 18. The (joint) differential entropy of a continuous random vector X n p X n(x) over R is: Z h(x

More information

Multimedia Communications. Mathematical Preliminaries for Lossless Compression

Multimedia Communications. Mathematical Preliminaries for Lossless Compression Multimedia Communications Mathematical Preliminaries for Lossless Compression What we will see in this chapter Definition of information and entropy Modeling a data source Definition of coding and when

More information

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy Coding and Information Theory Chris Williams, School of Informatics, University of Edinburgh Overview What is information theory? Entropy Coding Information Theory Shannon (1948): Information theory is

More information

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information 204 IEEE International Symposium on Information Theory Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information Omur Ozel, Kaya Tutuncuoglu 2, Sennur Ulukus, and Aylin Yener

More information

Example: Letter Frequencies

Example: Letter Frequencies Example: Letter Frequencies i a i p i 1 a 0.0575 2 b 0.0128 3 c 0.0263 4 d 0.0285 5 e 0.0913 6 f 0.0173 7 g 0.0133 8 h 0.0313 9 i 0.0599 10 j 0.0006 11 k 0.0084 12 l 0.0335 13 m 0.0235 14 n 0.0596 15 o

More information

Example: Letter Frequencies

Example: Letter Frequencies Example: Letter Frequencies i a i p i 1 a 0.0575 2 b 0.0128 3 c 0.0263 4 d 0.0285 5 e 0.0913 6 f 0.0173 7 g 0.0133 8 h 0.0313 9 i 0.0599 10 j 0.0006 11 k 0.0084 12 l 0.0335 13 m 0.0235 14 n 0.0596 15 o

More information

EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY

EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY Discrete Messages and Information Content, Concept of Amount of Information, Average information, Entropy, Information rate, Source coding to increase

More information

A One-to-One Code and Its Anti-Redundancy

A One-to-One Code and Its Anti-Redundancy A One-to-One Code and Its Anti-Redundancy W. Szpankowski Department of Computer Science, Purdue University July 4, 2005 This research is supported by NSF, NSA and NIH. Outline of the Talk. Prefix Codes

More information

(Classical) Information Theory II: Source coding

(Classical) Information Theory II: Source coding (Classical) Information Theory II: Source coding Sibasish Ghosh The Institute of Mathematical Sciences CIT Campus, Taramani, Chennai 600 113, India. p. 1 Abstract The information content of a random variable

More information

Shannon s A Mathematical Theory of Communication

Shannon s A Mathematical Theory of Communication Shannon s A Mathematical Theory of Communication Emre Telatar EPFL Kanpur October 19, 2016 First published in two parts in the July and October 1948 issues of BSTJ. First published in two parts in the

More information

MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016

MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016 MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016 Lecture 14: Information Theoretic Methods Lecturer: Jiaming Xu Scribe: Hilda Ibriga, Adarsh Barik, December 02, 2016 Outline f-divergence

More information

Distributed Lossless Compression. Distributed lossless compression system

Distributed Lossless Compression. Distributed lossless compression system Lecture #3 Distributed Lossless Compression (Reading: NIT 10.1 10.5, 4.4) Distributed lossless source coding Lossless source coding via random binning Time Sharing Achievability proof of the Slepian Wolf

More information