CHAPTER 3. P (B j A i ) P (B j ) =log 2. j=1

Size: px
Start display at page:

Download "CHAPTER 3. P (B j A i ) P (B j ) =log 2. j=1"

Transcription

1 CHAPTER 3 Problem 3. : Also : Hence : I(B j ; A i ) = log P (B j A i ) P (B j ) 4 P (B j )= P (B j,a i )= i= 3 P (A i )= P (B j,a i )= j= =log P (B j,a i ) P (B j )P (A i ).3, j=.7, j=.4, j=3.3, i=.7, i=.3, i=3., i=4. I(B ; A )=log =+.57 bits (.3)(.3).5 I(B ; A )=log =.76 bits (.3)(.7).5 I(B ; A 3 )=log =.943 bits (.3)(.3). I(B ; A 4 )=log =+.757 bits (.3)(.).8 I(B ; A )=log =.65 bits (.7)(.3).3 I(B ; A )=log =.64 bits (.7)(.7). I(B ; A 3 )=log =+.5 bits (.7)(.3).4 I(B ; A 4 )=log =.53 bits (.7)(.).3 I(B 3 ; A )=log =. bits (.4)(.3).9 I(B 3 ; A )=log =+.334 bits (.4)(.7) 6

2 .4 I(B 3 ; A 3 )=log =+.5 bits (.4)(.3).6 I(B 3 ; A 4 )=log =.556 bits (.4)(.) (b) The average mutual information will be : 3 4 I(B; A) = P (Ai, B j )I(B j ; A i )=.677 bits j= i= Problem 3. : H(B) = 3 j= P (B j )log P (B j ) = [.3 log log log.4] =.56 bits/letter Problem 3.3 : Let f(u) =u ln u. The first and second derivatives of f(u) are df du = u and d f du = >, u > u 7

3 Hence this function achieves its minimum at df = u =. The minimum value is f(u = du ) = so ln u = u, at u =. For all other values of u :<u<, u, we have f(u) > u > ln u. Problem 3.4 : We will show that I(X; Y ) I(X; Y ) = i j P (x i,y j )log P (x i,y j ) P (x i )P (y j ) = ln i j P (x i,y j )ln P (x i)p (y j ) P (x i,y j ) We use the inequality ln u u. We need only consider those terms for which P (x i,y j ) > ; then, applying the inequality to each term in I(X; Y ): I(X; Y ) ln i j P (x i,y j ) [ P (x i )P (y j ) P (x i,y j ] ) The first inequality becomes equality if and only if = ln i j [P (x i )P (y j ) P (x i,y j )] P (x i )P (y j ) P (x i,y j ) = P (x i )P (y j )=P (x i,y j ) when P (x i,y j ) >. Also, since the summations [P (x i )P (y j ) P (x i,y j )] i j contain only the terms for which P (x i,y j ) >, this term equals zero if and only if P (X i )P (Y j )=, when P (x i,y j )=. Therefore, both inequalitites become equalities and hence, I(X; Y )= if and only if X and Y are statistically independent. 8

4 Problem 3.5 : We shall prove that H(X) log n : H(X) log n = n i= p i log p i log n = n i= p i log p i n i= p i log n = n i= p i log np i = ni= p ln i ln np i ( ni= p ln i np i ) = Hence, H(X) log n. Also, if p i =/n i H(X) =logn. Problem 3.6 : By definition, the differential entropy is H(X) = p(x)logp(x)dx For the uniformly distributed random variable : a H(X) = a log dx =loga a (a) For a =, H(X) = (b) For a =4, H(X) = log 4 = log (c) For a =/4, H(X) =log 4 = log Problem 3.7 : (a) The following figure depicts the design of a ternary Huffman code (we follow the convention that the lower-probability branch is assigned a ) : 9

5 Codeword Probability æ (b) The average number of binary digits per source letter is : R = i P (x i )n i =(.45) + 3(.37) + 4(.8) + 5(.) =.83 bits/letter (c) The entropy of the source is : H(X) = i P (x i )logp(x i )=.8 bits/letter As it is expected the entropy of the source is less than the average length of each codeword. Problem 3.8 : The source entropy is : H(X) = 5 p i log =log5=.3 bits/letter i= p i (a) When we encode one letter at a time we require R = 3 bits/letter. Hence, the efficiency is.3/3 =.77 (77%).

6 (b) If we encode two letters at a time, we have 5 possible sequences. Hence, we need 5 bits per -letter symbol, or R =.5 bits/letter ; the efficiency is.3/.5 =.93. (c) In the case of encoding three letters at a time we have 5 possible sequences. Hence we need 7 bits per 3-letter symbol, so R =7/3 bits/letter; the efficiency is.3/(7/3) =.994. Problem 3.9 : (a) I(x i ; y j ) = log P (x i y j ) P (x i ) = log P (x i,y j ) P (x i )P (y j ) = log P (y j x i ) P (y j ) = log log P (y j ) P (y j x i ) = I(y j ) I(y j x i ) (b) I(x i ; y j ) = log P (x i y j ) P (x i ) = log P (x i,y j ) P (x i )P (y j ) = log +log log P (x i ) P (y j ) P (x i,y j ) = I(x i )+I(y j ) I(x i,y j ) Problem 3. : (a) H(X) = p( p) k log (p( p) k ) k= = p log (p) ( p) k p log ( p) (k )( p) k k= k= = p log (p) ( p) p log p ( p) ( ( p)) = log (p) p log p ( p)

7 (b) Clearly P (X = k X >K)=fork K. Ifk>K,then But, P (X = k X >K)= P (X = k, X > K) P (X >K) = p( p)k P (X >K) ( ) P (X >K) = p( p) k K = p ( p) k ( p) k k=k+ k= k= ( ) ( p)k = p =( p) K ( p) ( p) so that P (X = k X >K)= If we let k = K + l with l =,,...,then p( p)k ( p) K P (X = k X >K)= p( p)k ( p) l ( p) K = p( p) l that is P (X = k X > K) is the geometrically distributed. Hence, using the results of the first part we obtain H(X X >K) = p( p) l log (p( p) l ) l= = log (p) p log p ( p) Problem 3. : (a) The marginal distribution P (x) isgivenbyp (x) = y P (x, y). Hence, H(X) = P (x)logp(x) = P (x, y)logp (x) x x y = P (x, y)logp (x) x,y Similarly it is proved that H(Y )= x,y P (x, y)logp (y). (b) Using the inequality ln w w withw = P (x)p (y),weobtain P (x,y) ln P (x)p (y) P (x, y) P (x)p (y) P (x, y)

8 Multiplying the previous by P (x, y) and adding over x, y, weobtain P (x, y)lnp (x)p (y) P (x, y)lnp (x, y) P (x)p (y) P (x, y) = x,y x,y x,y x,y Hence, H(X, Y ) P (x, y)lnp (x)p (y) = P (x, y)(ln P (x)+lnp(y)) x,y x,y = P (x, y)lnp (x) P (x, y)lnp (y) =H(X)+H(Y ) x,y x,y Equality holds when P (x)p (y) P (x,y) =, i.e when X, Y are independent. (c) H(X, Y )=H(X)+H(Y X) =H(Y )+H(X Y) Also, from part (b), H(X, Y ) H(X)+H(Y ). Combining the two relations, we obtain H(Y )+H(X Y) H(X)+H(Y )= H(X Y ) H(X) Suppose now that the previous relation holds with equality. Then, x P (x)logp (x y) = x P (x)logp (x) x P (x) log( P (x) P (x y) )= However, P (x) is a lwa ys grea ter or equa l to P (x y), so that log(p (x)/p (x y)) is non-negative. Since P (x) >, the above equality holds if and only if log(p (x)/p (x y)) = or equivalently if and only if P (x)/p (x y) =. This implies that P (x y) =P (x) meaning that X and Y are independent. Problem 3. : The marginal probabilities are given by P (X =) = k P (X =,Y = k) =P (X =,Y =)+P (X =,Y =)= 3 P (X =) = k P (X =,Y = k) =P (X =,Y =)= 3 P (Y =) = k P (X = k, Y =)=P (X =,Y =)= 3 P (Y =) = k P (X = k, Y =)=P (X =,Y =)+P (X =,Y =)= 3 3

9 Hence, H(X) = P i log P i = ( i= 3 log log 3 )=.983 H(X) = P i log P i = ( i= 3 log log 3 )=.983 H(X, Y ) = i= 3 log 3 =.585 H(X Y ) = H(X, Y ) H(Y )= =.6667 H(Y X) = H(X, Y ) H(X) = =.6667 Problem 3.3 : H = lim H(X n X,...,X n ) n [ = lim ] P (x,...,x n )log n P (x n x,...,x n ) x,...,x n [ = lim ] P (x,...,x n )log n P (x n x n ) x,...,x n = lim P (x n,x n )log n P (x n x n ) x n,x n = lim H(X n X n ) n However, for a stationary process P (x n,x n )andp (x n x n ) are independent of n, sothat H = lim n H(X n X n )=H(X n X n ) Problem 3.4 : H(X, Y ) = H(X, g(x)) = H(X)+H(g(X) X) = H(g(X)) + H(X g(x)) 4

10 But, H(g(X) X) =, sinceg( ) is deterministic. Therefore, H(X) =H(g(X)) + H(X g(x)) Since each term in the previous equation is non-negative we obtain H(X) H(g(X)) Equality holds when H(X g(x)) =. This means that the values g(x) uniquely determine X, or that g( ) is aone to one mapping. Problem 3.5 : I(X; Y ) = n i= mj= P (x i,y j )log P (x i,y j ) P (x i )P (y j ) = = { ni= mj= P (x i,y j )logp (x i,y j ) n i= mj= P (x i,y j )logp (x i ) n i= mj= P (x i,y j )logp (y j ) { ni= mj= P (x i,y j )logp (x i,y j ) n i= P (x i )logp (x i ) m j= P (y j )logp (y j ) } } = H(XY )+H(X)+H(Y ) Problem 3.6 : H(X X...X n )= m m j = j =... m n j n= P (x,x,..., x n )logp (x,x,..., x n ) Since the {x i } are statistically independent : P (x,x,..., x n )=P(x )P (x )...P (x n ) and m m n... P (x )P (x )...P (x n )=P(x ) j = (similarly for the other x i ). Then : H(X X...X n ) = m j = j n= m j =... m n j n= P (x )P (x )...P (x n )logp (x )P (x )...P (x n ) = m j = P (x )logp (x ) m j = P (x )logp (x )... m n j n= P (x n )logp (x n ) = n i= H(X i ) 5

11 Problem 3.7 : We consider an n input, n output channel. Since it is noiseless : Hence : Butitisalsotruethat: Hence : P (y j x i )= {, i j, i=j H(X Y ) = n i= nj= P (x i,y j )logp (x i y j ) = n i= nj= P (y j x i )p(x i )logp (x i y j ) P (x i y j )= {, i j, i=j n H(X Y )= P (x i )log= i= } } Problem 3.8 : The conditional mutual information between x 3 and x given x is defined as : I(x 3 ; x x )=log P (x 3,x x ) P (x 3 x )P (x x ) =logp (x 3 x x ) P (x 3 x ) Hence : and I(x 3 ; x x )=I(x 3 x ) I(x 3 x x ) I(X 3 ; X X ) = x x x3 P (x,x,x 3 )log P (x 3 x x ) P (x 3 x ) = { x x x3 P (x,x,x 3 )logp(x 3 x ) + x x x3 P (x,x,x 3 )logp(x 3 x x ) } Since I(X 3 ; X X ), it follows that : = H(X 3 X ) H(X 3 X X ) H(X 3 X ) H(X 3 X X ) Problem 3.9 : 6

12 Assume that a>. Then we know that in the linear transformation Y = ax + b : Hence : p Y (y) = a p X( y b a ) H(Y ) = p Y (y)logp Y (y)dy = p a X( y b )log p a a X( y b )dy a Let u = y b. Then dy = adu, and : a H(Y ) = a p X(u)[logp X (u) log a] adu = p X(u)logp X (u)du + p X(u)logadu = H(X)+loga In a similar way, we can prove that for a<: H(Y )= H(X) log a Problem 3. : The linear transformation produces the symbols : y i = ax i + b, i =,, 3 with corresponding probabilities p =.45, p =.35, p 3 =.. since the {y i } have the same probability distribution as the {x i },itfollowsthat: H(Y )=H(X). Hence, the entropy of a DMS is not affected by the linear transformation. Problem 3. : (a) The following figure depicts the design of the Huffman code, when encoding a single level ata time: 7

13 Codeword Level Probability a.3365 a a a The average number of binary digits per source level is : æ R = i P (a i )n i =.995 bits/level The entropy of the source is : H(X) = i P (a i )logp(a i )=.98 bits/level (b) Encoding two levels at a time : 8

14 Codeword Levels a a Probability.33 a a a a.33 a a a a a a a a a a a 3 a a 3 a.55 a 4 a.55.4 a 4 a.55 a 3 a a 3 a a 4 a a 4 a æ The average number of binary digits per level pair is R = k P (a k )n k = bits/pair resulting in an average number : R =.937 bits/level (c) H(X) R J J <H(X)+ J As J, R J J H(X) =.98 bits/level. 9

15 Problem 3. : First, we need the state probabilities P (x i ), i =,. For stationary Markov processes, these can be found, in general, by the solution of the system : P Π=P, P i = where P is the state probability vector and Π is the transition matrix : Π[ij] = P (x j x i ). However, in the case of a two-state Markov source, we can find P (x i )inasimplerwaybynoting that the probability of a transition from state to state equals the probability of a transition from state to state (so that the probability of each state will remain the same). Hence : P (x x )P (x )=P (x x )P (x ).3P (x )=.P (x ) P (x )=.6, P(x )=.4 i Then : H(X) = { P (x )[ P (x x )logp (x x ) P (x x )logp (x x )] + P (x )[ P (x x )logp (x x ) P (x x )logp (x x )] } =.6[.8log.8.log.] +.4[.3log.3.7log.7] =.7857 bits/letter If the source is a binary DMS with output letter probabilities P (x )=.6, P(x )=.4, its entropy will be : H DMS (X) =.6log.6.4log.4 =.97 bits/letter We see that the entropy of the Markov source is smaller, since the memory inherent in it reduces the information content of each output. Problem 3.3 : (a) H(X) = (.5 log.5 +.log.+.log.+.5 log log log.5 +.3log.3) =.58 (b) After quantization, the new alphabet is B = { 4,, 4} and the corresponding symbol probabilities are given by P ( 4) = P ( 5) + P ( 3) =.5 +. =.5 P () = P ( ) + P () + P () = =.3 P (4) = P (3) + P (5) = =.55 3

16 Hence, H(Q(X)) =.46. As it is observed quantization decreases the entropy of the source. Problem 3.4 : The following figure depicts the design of a ternary Huffman code The average codeword length is R(X) = x P (x)n x =. + ( ) =.78 (ternary symbols/output) For a fair comparison of the average codeword length with the entropy of the source, we compute the latter with logarithms in base 3. Hence, H(X) = x P (x)log 3 P (x) =.747 As it is expected H(X) R(X). Problem 3.5 : Parsing the sequence by the rules of the Lempel-Ziv coding scheme we obtain the phrases,,,,,,,,,,,,,,,,,,... The number of the phrases is 8. For each phrase we need 5 bits plus an extra bit to represent the new source output. 3

17 Dictionary Dictionary Codeword Location Contents Problem 3.6 : (a) where we have used the fact (b) x H(X) = λ e λ ln( = ln( λ ) H(X) = x λ e x λ e x λ e λ )dx = lnλ + λ xdx λ = lnλ + λ λ =+lnλ λ dx + λ e x λ x λ dx λ e x λ dx =ande[x] = x λ e x λ dx = λ. = ln( λ ) x x λ e λ ln( λ e λ )dx x λ e λ dx + λ 3 x x λ e λ dx

18 [ = ln(λ)+ x λ λ e x λ dx + x = ln(λ)+ λ λ + λ λ =+ln(λ) x λ e λ dx ] (c) H(X) = x + λ ln λ λ = ( ) [ ln λ λ ( ) x + λ dx λ x + λ λ dx + x + λ ln(x + λ)dx λ λ = ln(λ ) λ z ln zdz λ = ln(λ ) λ [ z ln z ] λ z 4 λ λ ( ) x + λ x + λ ln dx λ λ ] x + λ dx λ λ x + λ λ ln( x + λ)dx = ln(λ ) ln(λ)+ Problem 3.7 : (a) Since R(D) =log λ and D = λ λ,weobtainr(d) =log( ) = log() = bit/sample. D λ/ (b) The following figure depicts R(D) for λ =.,. a nd.3. As it is observed from the figure, an increase of the parameter λ increases the required rate for a given distortion R(D) 4 3 l=.3 l=. l= Distortion D 33

19 Problem 3.8 : (a) For a Gaussian random variable of zero mean and variance σ the rate-distortion function is given by R(D) = log σ. Hence, the upper bound is satisfied with equality. For the lower D bound recall that H(X) = log (πeσ ). Thus, H(X) log (πed) = log (πeσ ) log (πed) = ( ) πeσ log = R(D) πed As it is observed the upper and the lower bounds coincide. (b) The differential entropy of a Laplacian source with parameter λ is H(X) = + ln(λ). The variance of the Laplacian distribution is σ = x x λ e λ dx =λ Hence, with σ =,weobtainλ = /andh(x) =+ln(λ) =+ln( ) =.3466 nats/symbol =.5 bits/symbol. A plot of the lower and upper bound of R(D) is given in the next figure. 5 Laplacian Distribution, unit variance 4 3 R(D) Upper Bound Lower Bound Distortion D (c) The variance of the triangular distribution is given by ( ) ( ) x + λ λ x + λ σ = x dx + x dx λ λ = λ ( 4 x4 + λ 3 x3 ) = λ 6 λ 34 λ + λ ( 4 x4 + λ 3 x3 ) λ

20 Hence, with σ =,weobtainλ = 6andH(X) =ln(6) ln( 6)+/ =.795 bits /source output. A plot of the lower and upper bound of R(D) is given in the next figure. 4.5 Triangular distribution, unit variance R(D).5.5 Lower Bound Upper Bound Distortion D Problem 3.9 : σ = E[X (t)] = R X (τ) τ= = A Hence, SQNR = 3 4 ν X =3 4 ν X =3 4 ν A x max A With SQNR = 6 db, we obtain ( 3 4 q ) log =6= q = The smallest integer larger that q is. Hence, the required number of quantization levels is ν =. Problem 3.3 : (a) H(X G) = p(x, g)logp(x g)dxdg 35

21 But X, G are independent, so : p(x, g) = p(x)p(g), p(x g) = p(x).hence : H(X G) = p(g) [ p(x)logp(x)dx] dg = p(g)h(x)dg = H(X) = log(πeσ x ) where the last equality stems from the Gaussian pdf of X. (b) I(X; Y )=H(Y ) H(Y X) Since Y is the sum of two independent, zero-mean Gaussian r.v s, it is also a zero-mean Gaussian r.v. with variance : σ y = σ x + σ n. Hence : H(Y )= log (πe (σ x + σ n )). Also, since y = x + g : p(y x) =p g (y x) = e (y x) σn πσn Hence : H(Y X) = p(x, y)logp(y x)dxdy ( ) (y x) = p(x)loge p(y x)ln exp( ) dydx πσn σn [ ( = p(x)loge p g (y x) ln( ) ] (y x) πσ n )+ dy dx σn [ = p(x)loge ln( πσ n )+ ] σ σn n dx = [log( πσ n )+ ] log e p(x)dx = log ( πeσ n) (= H(G)) where we have used the fact that : p g(y x)dy =, (y x) p g (y x)dy = E [G ]=σn. From H(Y ), H(Y X) : I(X; Y )=H(Y) H(Y X) = log ( πe(σx + σ n )) log ( ( ) ) πeσn = log + σ x σn 36

22 Problem 3.3 : Codeword Letter Probability x.5 x. x x 4. x 5. x 6.8 x 7 x 8 x æ R =.85 ternary symbols/letter Problem 3.3 : Given (n,n,n 3,n 4 )=(,,, 3) we have : 4 n k = = 9 k= 8 > Since the Craft inequality is not satisfied, a binary code with code word lengths (,,, 3) that satisfies the prefix condition does not exist. Problem 3.33 : n n k = n k= k= n = n n = 37

23 Therefore the Kraft inequality is satisfied. Problem 3.34 : But : and Hence : p(x) = (π) n/ M H(X) =... log p(x) = log(π)n M e X M X / p(x)logp(x)dx ( log e ) X M X ( )... log e X M X p(x)dx = n log e H(X) = log(π)n M + log en = log(πe)n M Problem 3.35 : R(D) =+D log D +( D)log( D), D = P e / R(D) D 38

24 Problem 3.36 : R(D) =logm + D log D +( D)log ( D) M 3.5 M=8 R(D).5 M=4.5 M= D Problem 3.37 : Let W = P P. Then : d W (X, X) =(X X) W(X X) d W (X, X) = (X X) P P(X X) = ( P(X X) ) P(X X) ( ) ( Y Ỹ Y Ỹ ) = n where by definition : Y= npx, Ỹ = np X. Hence : d W (X, X) =d (Y,Ỹ). Problem 3.38 : (a) The first order predictor is : ˆx(n) =a x(n ). The coefficient a that minimizes the MSE is found from the orthogonality of the prediction error to the prediction data : E [e(n)x(n )] = E [(x(n) a x(n )) x(n )] = φ() a φ() = a = φ()/φ() = / 39

25 The minimum MSE is : ɛ = φ() ( a )=3/4 (b) For the second order predictor : ˆx(n) =a x(n ) + a x(n ). Following the Levinson- Durbin algorithm (Eqs 3-5-5) : a = φ() k= a k φ( k) = ɛ 3/4 a = a a a =/3 = /3 The minimum MSE is : ɛ = ɛ ( a ) =/3 Problem 3.39 : p(x,x )= { 5 7ab, x,x C, o.w If x,x are quantized separately by using uniform intervals of length, the number of levels needed is L = a,l = b. The number of bits is : R x = R + R =logl +logl =log ab By using vector quantization with squares having area,wehavel x = 7ab and R 5 x =logl x = log 7ab bits. The difference in bit rate is : 5 R x R x =log ab 7ab log 5 =log5 =. bits/output sample 7 for all a, b >. } Problem 3.4 : (a) The area between the two squares is 4 4 =. Hence, p X,Y (x, y) =. The marginal probability p X (x) isgivenbyp X (x) = p X,Y (x, y)dy. If X<, then p X (x) = p X,Y (x, y)dy = y = 3 4

26 If X<, then Finally, if X, then p X (x) = dy + dy = 6 p X (x) = p X,Y (x, y)dy = y The next figure depicts the marginal distribution p X (x).... /3 /6 = 3 Similarly we find that - - y< 3 p Y (y) = y< 6 y 3 (b) The quantization levels ˆx,ˆx,ˆx 3 and ˆx 4 are set to 3,, and 3 resulting distortion is respectively. The D X = (x + 3 ) p X (x)dx + = (x +3x )dx + 6 ( 3 x3 + 3 x + 9 ) 4 x (x + ) p X (x)dx (x + x + 4 )dx ( 3 x3 + x + 4 x ) The total distortion is = 3 = + 6 D total = D X + D Y = + = 6 whereas the resulting number of bits per (X, Y )pair R = R X + R Y =log 4+log 4=4 (c) Suppose that we divide the region over which p(x, y) intol equal subregions. The case of L = 4 is depicted in the next figure. 4

27 For each subregion the quantization output vector (ˆx, ŷ) is the centroid of the corresponding rectangle. Since, each subregion has the same shape (uniform quantization), a rectangle with width equal to one and length /L, the distortion of the vector quantizer is D = = L = L L [ L [(x, y) (, L L )] dxdy [ (x ) +(y L + 3 L 3 ] L ) = + L ] dxdy If we set D = 6,weobtain L = = L = 44 = Thus, we have to divide the area over which p(x, y), into equal subregions in order to achieve the same distortion. In this case the resulting number of bits per source output pair (X, Y )isr =log = Problem 3.4 : (a) The joint probability density function is p XY (x, y) = ( =. The marginal distribution ) 8 p X (x) isp X (x) = y p XY (x, y)dy. If x,then If x,then p X (x) = p X (x) = The next figure depicts p X (x). x+ x x+ x p X,Y (x, y)dy = 8 y x+ x = x + 4 p X,Y (x, y)dy = 8 y x+ x = x + 4 4

28 From the symmetry of the problem we have p Y (y) = { y+ 4 y< y+ 4 y (b) D X = (x + 3 ) p X (x)dx + (x + ) p X (x)dx = (x + 3 ) (x +)dx + (x + ) ( x +)dx ( 4 x x x + 9 ) x ( 4 x4 + x x + ) x = = The total distortion is + D total = D X + D Y = + = 6 whereas the required number of bits per source output pair R = R X + R Y =log 4+log 4=4 (c) We divide the square over which p(x, y) into 4 = 6 equal square regions. The area of each square is and the resulting distortion D = 6 [ (x 8 ) +(y ] ) dxdy = 4 (x ) dxdy = 4 (x + 8 x )dx = 4 ( 3 x3 + 8 x ) x = Hence, using vector quantization and the same rate we obtain half the distortion. 43

ECE 4400:693 - Information Theory

ECE 4400:693 - Information Theory ECE 4400:693 - Information Theory Dr. Nghi Tran Lecture 8: Differential Entropy Dr. Nghi Tran (ECE-University of Akron) ECE 4400:693 Lecture 1 / 43 Outline 1 Review: Entropy of discrete RVs 2 Differential

More information

Coding for Discrete Source

Coding for Discrete Source EGR 544 Communication Theory 3. Coding for Discrete Sources Z. Aliyazicioglu Electrical and Computer Engineering Department Cal Poly Pomona Coding for Discrete Source Coding Represent source data effectively

More information

Lecture 8: Channel Capacity, Continuous Random Variables

Lecture 8: Channel Capacity, Continuous Random Variables EE376A/STATS376A Information Theory Lecture 8-02/0/208 Lecture 8: Channel Capacity, Continuous Random Variables Lecturer: Tsachy Weissman Scribe: Augustine Chemparathy, Adithya Ganesh, Philip Hwang Channel

More information

Exercises with solutions (Set B)

Exercises with solutions (Set B) Exercises with solutions (Set B) 3. A fair coin is tossed an infinite number of times. Let Y n be a random variable, with n Z, that describes the outcome of the n-th coin toss. If the outcome of the n-th

More information

Solutions to Homework Set #1 Sanov s Theorem, Rate distortion

Solutions to Homework Set #1 Sanov s Theorem, Rate distortion st Semester 00/ Solutions to Homework Set # Sanov s Theorem, Rate distortion. Sanov s theorem: Prove the simple version of Sanov s theorem for the binary random variables, i.e., let X,X,...,X n be a sequence

More information

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1 Kraft s inequality An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if N 2 l i 1 Proof: Suppose that we have a tree code. Let l max = max{l 1,...,

More information

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions Engineering Tripos Part IIA THIRD YEAR 3F: Signals and Systems INFORMATION THEORY Examples Paper Solutions. Let the joint probability mass function of two binary random variables X and Y be given in the

More information

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University Chapter 4 Data Transmission and Channel Capacity Po-Ning Chen, Professor Department of Communications Engineering National Chiao Tung University Hsin Chu, Taiwan 30050, R.O.C. Principle of Data Transmission

More information

EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY

EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY Discrete Messages and Information Content, Concept of Amount of Information, Average information, Entropy, Information rate, Source coding to increase

More information

Chapter 2: Source coding

Chapter 2: Source coding Chapter 2: meghdadi@ensil.unilim.fr University of Limoges Chapter 2: Entropy of Markov Source Chapter 2: Entropy of Markov Source Markov model for information sources Given the present, the future is independent

More information

Lecture 17: Differential Entropy

Lecture 17: Differential Entropy Lecture 17: Differential Entropy Differential entropy AEP for differential entropy Quantization Maximum differential entropy Estimation counterpart of Fano s inequality Dr. Yao Xie, ECE587, Information

More information

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions EE/Stat 376B Handout #5 Network Information Theory October, 14, 014 1. Problem.4 parts (b) and (c). Homework Set # Solutions (b) Consider h(x + Y ) h(x + Y Y ) = h(x Y ) = h(x). (c) Let ay = Y 1 + Y, where

More information

Exercises with solutions (Set D)

Exercises with solutions (Set D) Exercises with solutions Set D. A fair die is rolled at the same time as a fair coin is tossed. Let A be the number on the upper surface of the die and let B describe the outcome of the coin toss, where

More information

EE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm

EE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm EE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm 1. Feedback does not increase the capacity. Consider a channel with feedback. We assume that all the recieved outputs are sent back immediately

More information

Chapter I: Fundamental Information Theory

Chapter I: Fundamental Information Theory ECE-S622/T62 Notes Chapter I: Fundamental Information Theory Ruifeng Zhang Dept. of Electrical & Computer Eng. Drexel University. Information Source Information is the outcome of some physical processes.

More information

Lecture 2. Capacity of the Gaussian channel

Lecture 2. Capacity of the Gaussian channel Spring, 207 5237S, Wireless Communications II 2. Lecture 2 Capacity of the Gaussian channel Review on basic concepts in inf. theory ( Cover&Thomas: Elements of Inf. Theory, Tse&Viswanath: Appendix B) AWGN

More information

EE376A: Homeworks #4 Solutions Due on Thursday, February 22, 2018 Please submit on Gradescope. Start every question on a new page.

EE376A: Homeworks #4 Solutions Due on Thursday, February 22, 2018 Please submit on Gradescope. Start every question on a new page. EE376A: Homeworks #4 Solutions Due on Thursday, February 22, 28 Please submit on Gradescope. Start every question on a new page.. Maximum Differential Entropy (a) Show that among all distributions supported

More information

Ch. 8 Math Preliminaries for Lossy Coding. 8.4 Info Theory Revisited

Ch. 8 Math Preliminaries for Lossy Coding. 8.4 Info Theory Revisited Ch. 8 Math Preliminaries for Lossy Coding 8.4 Info Theory Revisited 1 Info Theory Goals for Lossy Coding Again just as for the lossless case Info Theory provides: Basis for Algorithms & Bounds on Performance

More information

CS6304 / Analog and Digital Communication UNIT IV - SOURCE AND ERROR CONTROL CODING PART A 1. What is the use of error control coding? The main use of error control coding is to reduce the overall probability

More information

18.2 Continuous Alphabet (discrete-time, memoryless) Channel

18.2 Continuous Alphabet (discrete-time, memoryless) Channel 0-704: Information Processing and Learning Spring 0 Lecture 8: Gaussian channel, Parallel channels and Rate-distortion theory Lecturer: Aarti Singh Scribe: Danai Koutra Disclaimer: These notes have not

More information

ECE Information theory Final

ECE Information theory Final ECE 776 - Information theory Final Q1 (1 point) We would like to compress a Gaussian source with zero mean and variance 1 We consider two strategies In the first, we quantize with a step size so that the

More information

Information Theory. Lecture 5 Entropy rate and Markov sources STEFAN HÖST

Information Theory. Lecture 5 Entropy rate and Markov sources STEFAN HÖST Information Theory Lecture 5 Entropy rate and Markov sources STEFAN HÖST Universal Source Coding Huffman coding is optimal, what is the problem? In the previous coding schemes (Huffman and Shannon-Fano)it

More information

COMM901 Source Coding and Compression. Quiz 1

COMM901 Source Coding and Compression. Quiz 1 German University in Cairo - GUC Faculty of Information Engineering & Technology - IET Department of Communication Engineering Winter Semester 2013/2014 Students Name: Students ID: COMM901 Source Coding

More information

EE376A - Information Theory Final, Monday March 14th 2016 Solutions. Please start answering each question on a new page of the answer booklet.

EE376A - Information Theory Final, Monday March 14th 2016 Solutions. Please start answering each question on a new page of the answer booklet. EE376A - Information Theory Final, Monday March 14th 216 Solutions Instructions: You have three hours, 3.3PM - 6.3PM The exam has 4 questions, totaling 12 points. Please start answering each question on

More information

Lecture 22: Final Review

Lecture 22: Final Review Lecture 22: Final Review Nuts and bolts Fundamental questions and limits Tools Practical algorithms Future topics Dr Yao Xie, ECE587, Information Theory, Duke University Basics Dr Yao Xie, ECE587, Information

More information

Revision of Lecture 5

Revision of Lecture 5 Revision of Lecture 5 Information transferring across channels Channel characteristics and binary symmetric channel Average mutual information Average mutual information tells us what happens to information

More information

Motivation for Arithmetic Coding

Motivation for Arithmetic Coding Motivation for Arithmetic Coding Motivations for arithmetic coding: 1) Huffman coding algorithm can generate prefix codes with a minimum average codeword length. But this length is usually strictly greater

More information

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy Coding and Information Theory Chris Williams, School of Informatics, University of Edinburgh Overview What is information theory? Entropy Coding Information Theory Shannon (1948): Information theory is

More information

TSBK08 Data compression Exercises

TSBK08 Data compression Exercises TSBK08 Data compression Exercises Contents 1 Information theory............................... 1 2 Source coding.................................. 5 3 Differential entropy and rate-distortion theory................

More information

Electrical and Information Technology. Information Theory. Problems and Solutions. Contents. Problems... 1 Solutions...7

Electrical and Information Technology. Information Theory. Problems and Solutions. Contents. Problems... 1 Solutions...7 Electrical and Information Technology Information Theory Problems and Solutions Contents Problems.......... Solutions...........7 Problems 3. In Problem?? the binomial coefficent was estimated with Stirling

More information

4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information

4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information 4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information Ramji Venkataramanan Signal Processing and Communications Lab Department of Engineering ramji.v@eng.cam.ac.uk

More information

Lecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157

Lecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157 Lecture 6: Gaussian Channels Copyright G. Caire (Sample Lectures) 157 Differential entropy (1) Definition 18. The (joint) differential entropy of a continuous random vector X n p X n(x) over R is: Z h(x

More information

EE4601 Communication Systems

EE4601 Communication Systems EE4601 Communication Systems Week 2 Review of Probability, Important Distributions 0 c 2011, Georgia Institute of Technology (lect2 1) Conditional Probability Consider a sample space that consists of two

More information

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria Source Coding Master Universitario en Ingeniería de Telecomunicación I. Santamaría Universidad de Cantabria Contents Introduction Asymptotic Equipartition Property Optimal Codes (Huffman Coding) Universal

More information

A One-to-One Code and Its Anti-Redundancy

A One-to-One Code and Its Anti-Redundancy A One-to-One Code and Its Anti-Redundancy W. Szpankowski Department of Computer Science, Purdue University July 4, 2005 This research is supported by NSF, NSA and NIH. Outline of the Talk. Prefix Codes

More information

PROOF OF ZADOR-GERSHO THEOREM

PROOF OF ZADOR-GERSHO THEOREM ZADOR-GERSHO THEOREM FOR VARIABLE-RATE VQ For a stationary source and large R, the least distortion of k-dim'l VQ with nth-order entropy coding and rate R or less is δ(k,n,r) m k * σ 2 η kn 2-2R = Z(k,n,R)

More information

Chapter 3 Source Coding. 3.1 An Introduction to Source Coding 3.2 Optimal Source Codes 3.3 Shannon-Fano Code 3.4 Huffman Code

Chapter 3 Source Coding. 3.1 An Introduction to Source Coding 3.2 Optimal Source Codes 3.3 Shannon-Fano Code 3.4 Huffman Code Chapter 3 Source Coding 3. An Introduction to Source Coding 3.2 Optimal Source Codes 3.3 Shannon-Fano Code 3.4 Huffman Code 3. An Introduction to Source Coding Entropy (in bits per symbol) implies in average

More information

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye Chapter 2: Entropy and Mutual Information Chapter 2 outline Definitions Entropy Joint entropy, conditional entropy Relative entropy, mutual information Chain rules Jensen s inequality Log-sum inequality

More information

Chapter 9 Fundamental Limits in Information Theory

Chapter 9 Fundamental Limits in Information Theory Chapter 9 Fundamental Limits in Information Theory Information Theory is the fundamental theory behind information manipulation, including data compression and data transmission. 9.1 Introduction o For

More information

LECTURE 3. Last time:

LECTURE 3. Last time: LECTURE 3 Last time: Mutual Information. Convexity and concavity Jensen s inequality Information Inequality Data processing theorem Fano s Inequality Lecture outline Stochastic processes, Entropy rate

More information

Chapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye

Chapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye Chapter 8: Differential entropy Chapter 8 outline Motivation Definitions Relation to discrete entropy Joint and conditional differential entropy Relative entropy and mutual information Properties AEP for

More information

Chapter 2 Date Compression: Source Coding. 2.1 An Introduction to Source Coding 2.2 Optimal Source Codes 2.3 Huffman Code

Chapter 2 Date Compression: Source Coding. 2.1 An Introduction to Source Coding 2.2 Optimal Source Codes 2.3 Huffman Code Chapter 2 Date Compression: Source Coding 2.1 An Introduction to Source Coding 2.2 Optimal Source Codes 2.3 Huffman Code 2.1 An Introduction to Source Coding Source coding can be seen as an efficient way

More information

MARKOV CHAINS A finite state Markov chain is a sequence of discrete cv s from a finite alphabet where is a pmf on and for

MARKOV CHAINS A finite state Markov chain is a sequence of discrete cv s from a finite alphabet where is a pmf on and for MARKOV CHAINS A finite state Markov chain is a sequence S 0,S 1,... of discrete cv s from a finite alphabet S where q 0 (s) is a pmf on S 0 and for n 1, Q(s s ) = Pr(S n =s S n 1 =s ) = Pr(S n =s S n 1

More information

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real

More information

Ch. 8 Math Preliminaries for Lossy Coding. 8.5 Rate-Distortion Theory

Ch. 8 Math Preliminaries for Lossy Coding. 8.5 Rate-Distortion Theory Ch. 8 Math Preliminaries for Lossy Coding 8.5 Rate-Distortion Theory 1 Introduction Theory provide insight into the trade between Rate & Distortion This theory is needed to answer: What do typical R-D

More information

PART III. Outline. Codes and Cryptography. Sources. Optimal Codes (I) Jorge L. Villar. MAMME, Fall 2015

PART III. Outline. Codes and Cryptography. Sources. Optimal Codes (I) Jorge L. Villar. MAMME, Fall 2015 Outline Codes and Cryptography 1 Information Sources and Optimal Codes 2 Building Optimal Codes: Huffman Codes MAMME, Fall 2015 3 Shannon Entropy and Mutual Information PART III Sources Information source:

More information

at Some sort of quantization is necessary to represent continuous signals in digital form

at Some sort of quantization is necessary to represent continuous signals in digital form Quantization at Some sort of quantization is necessary to represent continuous signals in digital form x(n 1,n ) x(t 1,tt ) D Sampler Quantizer x q (n 1,nn ) Digitizer (A/D) Quantization is also used for

More information

Digital Image Processing Lectures 25 & 26

Digital Image Processing Lectures 25 & 26 Lectures 25 & 26, Professor Department of Electrical and Computer Engineering Colorado State University Spring 2015 Area 4: Image Encoding and Compression Goal: To exploit the redundancies in the image

More information

ECE353: Probability and Random Processes. Lecture 7 -Continuous Random Variable

ECE353: Probability and Random Processes. Lecture 7 -Continuous Random Variable ECE353: Probability and Random Processes Lecture 7 -Continuous Random Variable Xiao Fu School of Electrical Engineering and Computer Science Oregon State University E-mail: xiao.fu@oregonstate.edu Continuous

More information

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows. Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage

More information

Continuous Random Variables

Continuous Random Variables 1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables

More information

ELEC546 Review of Information Theory

ELEC546 Review of Information Theory ELEC546 Review of Information Theory Vincent Lau 1/1/004 1 Review of Information Theory Entropy: Measure of uncertainty of a random variable X. The entropy of X, H(X), is given by: If X is a discrete random

More information

1 Introduction to information theory

1 Introduction to information theory 1 Introduction to information theory 1.1 Introduction In this chapter we present some of the basic concepts of information theory. The situations we have in mind involve the exchange of information through

More information

MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006)

MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006) MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK SATELLITE COMMUNICATION DEPT./SEM.:ECE/VIII UNIT V PART-A 1. What is binary symmetric channel (AUC DEC 2006) 2. Define information rate? (AUC DEC 2007)

More information

Introduction to Estimation and Data fusion Part I: Probability, State and Information Models

Introduction to Estimation and Data fusion Part I: Probability, State and Information Models Introduction to Estimation and Data fusion Part I: Probability, State and Information Models Hugh Durrant-Whyte ARC Centre of Excellence for Autonomous Systems Australian Centre for Field Robotics The

More information

ELEMENT OF INFORMATION THEORY

ELEMENT OF INFORMATION THEORY History Table of Content ELEMENT OF INFORMATION THEORY O. Le Meur olemeur@irisa.fr Univ. of Rennes 1 http://www.irisa.fr/temics/staff/lemeur/ October 2010 1 History Table of Content VERSION: 2009-2010:

More information

Solutions to Homework Set #4 Differential Entropy and Gaussian Channel

Solutions to Homework Set #4 Differential Entropy and Gaussian Channel Solutions to Homework Set #4 Differential Entropy and Gaussian Channel 1. Differential entropy. Evaluate the differential entropy h(x = f lnf for the following: (a Find the entropy of the exponential density

More information

Lecture 3. Mathematical methods in communication I. REMINDER. A. Convex Set. A set R is a convex set iff, x 1,x 2 R, θ, 0 θ 1, θx 1 + θx 2 R, (1)

Lecture 3. Mathematical methods in communication I. REMINDER. A. Convex Set. A set R is a convex set iff, x 1,x 2 R, θ, 0 θ 1, θx 1 + θx 2 R, (1) 3- Mathematical methods in communication Lecture 3 Lecturer: Haim Permuter Scribe: Yuval Carmel, Dima Khaykin, Ziv Goldfeld I. REMINDER A. Convex Set A set R is a convex set iff, x,x 2 R, θ, θ, θx + θx

More information

Introduction to Probability and Stocastic Processes - Part I

Introduction to Probability and Stocastic Processes - Part I Introduction to Probability and Stocastic Processes - Part I Lecture 1 Henrik Vie Christensen vie@control.auc.dk Department of Control Engineering Institute of Electronic Systems Aalborg University Denmark

More information

conditional cdf, conditional pdf, total probability theorem?

conditional cdf, conditional pdf, total probability theorem? 6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random

More information

MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK. SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A

MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK. SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK DEPARTMENT: ECE SEMESTER: IV SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A 1. What is binary symmetric channel (AUC DEC

More information

3. Probability and Statistics

3. Probability and Statistics FE661 - Statistical Methods for Financial Engineering 3. Probability and Statistics Jitkomut Songsiri definitions, probability measures conditional expectations correlation and covariance some important

More information

Lecture 11: Continuous-valued signals and differential entropy

Lecture 11: Continuous-valued signals and differential entropy Lecture 11: Continuous-valued signals and differential entropy Biology 429 Carl Bergstrom September 20, 2008 Sources: Parts of today s lecture follow Chapter 8 from Cover and Thomas (2007). Some components

More information

Lecture 5 Channel Coding over Continuous Channels

Lecture 5 Channel Coding over Continuous Channels Lecture 5 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 14, 2014 1 / 34 I-Hsiang Wang NIT Lecture 5 From

More information

2 Functions of random variables

2 Functions of random variables 2 Functions of random variables A basic statistical model for sample data is a collection of random variables X 1,..., X n. The data are summarised in terms of certain sample statistics, calculated as

More information

SOURCE CODING WITH SIDE INFORMATION AT THE DECODER (WYNER-ZIV CODING) FEB 13, 2003

SOURCE CODING WITH SIDE INFORMATION AT THE DECODER (WYNER-ZIV CODING) FEB 13, 2003 SOURCE CODING WITH SIDE INFORMATION AT THE DECODER (WYNER-ZIV CODING) FEB 13, 2003 SLEPIAN-WOLF RESULT { X i} RATE R x ENCODER 1 DECODER X i V i {, } { V i} ENCODER 0 RATE R v Problem: Determine R, the

More information

Homework 1 Due: Thursday 2/5/2015. Instructions: Turn in your homework in class on Thursday 2/5/2015

Homework 1 Due: Thursday 2/5/2015. Instructions: Turn in your homework in class on Thursday 2/5/2015 10-704 Homework 1 Due: Thursday 2/5/2015 Instructions: Turn in your homework in class on Thursday 2/5/2015 1. Information Theory Basics and Inequalities C&T 2.47, 2.29 (a) A deck of n cards in order 1,

More information

Lecture 5: Asymptotic Equipartition Property

Lecture 5: Asymptotic Equipartition Property Lecture 5: Asymptotic Equipartition Property Law of large number for product of random variables AEP and consequences Dr. Yao Xie, ECE587, Information Theory, Duke University Stock market Initial investment

More information

Chapter 2: Random Variables

Chapter 2: Random Variables ECE54: Stochastic Signals and Systems Fall 28 Lecture 2 - September 3, 28 Dr. Salim El Rouayheb Scribe: Peiwen Tian, Lu Liu, Ghadir Ayache Chapter 2: Random Variables Example. Tossing a fair coin twice:

More information

Review: mostly probability and some statistics

Review: mostly probability and some statistics Review: mostly probability and some statistics C2 1 Content robability (should know already) Axioms and properties Conditional probability and independence Law of Total probability and Bayes theorem Random

More information

EGR 544 Communication Theory

EGR 544 Communication Theory EGR 544 Communcaton Theory. Informaton Sources Z. Alyazcoglu Electrcal and Computer Engneerng Department Cal Poly Pomona Introducton Informaton Source x n Informaton sources Analog sources Dscrete sources

More information

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16 EE539R: Problem Set 4 Assigned: 3/08/6, Due: 07/09/6. Cover and Thomas: Problem 3.5 Sets defined by probabilities: Define the set C n (t = {x n : P X n(x n 2 nt } (a We have = P X n(x n P X n(x n 2 nt

More information

Multimedia Communications. Mathematical Preliminaries for Lossless Compression

Multimedia Communications. Mathematical Preliminaries for Lossless Compression Multimedia Communications Mathematical Preliminaries for Lossless Compression What we will see in this chapter Definition of information and entropy Modeling a data source Definition of coding and when

More information

Information Dimension

Information Dimension Information Dimension Mina Karzand Massachusetts Institute of Technology November 16, 2011 1 / 26 2 / 26 Let X would be a real-valued random variable. For m N, the m point uniform quantized version of

More information

Quiz 2 Date: Monday, November 21, 2016

Quiz 2 Date: Monday, November 21, 2016 10-704 Information Processing and Learning Fall 2016 Quiz 2 Date: Monday, November 21, 2016 Name: Andrew ID: Department: Guidelines: 1. PLEASE DO NOT TURN THIS PAGE UNTIL INSTRUCTED. 2. Write your name,

More information

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

Lecture 6 I. CHANNEL CODING. X n (m) P Y X 6- Introduction to Information Theory Lecture 6 Lecturer: Haim Permuter Scribe: Yoav Eisenberg and Yakov Miron I. CHANNEL CODING We consider the following channel coding problem: m = {,2,..,2 nr} Encoder

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Coding of memoryless sources 1/35

Coding of memoryless sources 1/35 Coding of memoryless sources 1/35 Outline 1. Morse coding ; 2. Definitions : encoding, encoding efficiency ; 3. fixed length codes, encoding integers ; 4. prefix condition ; 5. Kraft and Mac Millan theorems

More information

Problem 7.7 : We assume that P (x i )=1/3, i =1, 2, 3. Then P (y 1 )= 1 ((1 p)+p) = P (y j )=1/3, j=2, 3. Hence : and similarly.

Problem 7.7 : We assume that P (x i )=1/3, i =1, 2, 3. Then P (y 1 )= 1 ((1 p)+p) = P (y j )=1/3, j=2, 3. Hence : and similarly. (b) We note that the above capacity is the same to the capacity of the binary symmetric channel. Indeed, if we considerthe grouping of the output symbols into a = {y 1,y 2 } and b = {y 3,y 4 } we get a

More information

Information Theory CHAPTER. 5.1 Introduction. 5.2 Entropy

Information Theory CHAPTER. 5.1 Introduction. 5.2 Entropy Haykin_ch05_pp3.fm Page 207 Monday, November 26, 202 2:44 PM CHAPTER 5 Information Theory 5. Introduction As mentioned in Chapter and reiterated along the way, the purpose of a communication system is

More information

EECS 229A Spring 2007 * * (a) By stationarity and the chain rule for entropy, we have

EECS 229A Spring 2007 * * (a) By stationarity and the chain rule for entropy, we have EECS 229A Spring 2007 * * Solutions to Homework 3 1. Problem 4.11 on pg. 93 of the text. Stationary processes (a) By stationarity and the chain rule for entropy, we have H(X 0 ) + H(X n X 0 ) = H(X 0,

More information

Basic Principles of Video Coding

Basic Principles of Video Coding Basic Principles of Video Coding Introduction Categories of Video Coding Schemes Information Theory Overview of Video Coding Techniques Predictive coding Transform coding Quantization Entropy coding Motion

More information

lossless, optimal compressor

lossless, optimal compressor 6. Variable-length Lossless Compression The principal engineering goal of compression is to represent a given sequence a, a 2,..., a n produced by a source as a sequence of bits of minimal possible length.

More information

Channel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5.

Channel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5. Channel capacity Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5. Exercices Exercise session 11 : Channel capacity 1 1. Source entropy Given X a memoryless

More information

1 Random Variable: Topics

1 Random Variable: Topics Note: Handouts DO NOT replace the book. In most cases, they only provide a guideline on topics and an intuitive feel. 1 Random Variable: Topics Chap 2, 2.1-2.4 and Chap 3, 3.1-3.3 What is a random variable?

More information

Gaussian source Assumptions d = (x-y) 2, given D, find lower bound of I(X;Y)

Gaussian source Assumptions d = (x-y) 2, given D, find lower bound of I(X;Y) Gaussian source Assumptions d = (x-y) 2, given D, find lower bound of I(X;Y) E{(X-Y) 2 } D

More information

Information and Entropy

Information and Entropy Information and Entropy Shannon s Separation Principle Source Coding Principles Entropy Variable Length Codes Huffman Codes Joint Sources Arithmetic Codes Adaptive Codes Thomas Wiegand: Digital Image Communication

More information

Basic Principles of Lossless Coding. Universal Lossless coding. Lempel-Ziv Coding. 2. Exploit dependences between successive symbols.

Basic Principles of Lossless Coding. Universal Lossless coding. Lempel-Ziv Coding. 2. Exploit dependences between successive symbols. Universal Lossless coding Lempel-Ziv Coding Basic principles of lossless compression Historical review Variable-length-to-block coding Lempel-Ziv coding 1 Basic Principles of Lossless Coding 1. Exploit

More information

3F1 Information Theory, Lecture 3

3F1 Information Theory, Lecture 3 3F1 Information Theory, Lecture 3 Jossy Sayir Department of Engineering Michaelmas 2011, 28 November 2011 Memoryless Sources Arithmetic Coding Sources with Memory 2 / 19 Summary of last lecture Prefix-free

More information

EE376A - Information Theory Midterm, Tuesday February 10th. Please start answering each question on a new page of the answer booklet.

EE376A - Information Theory Midterm, Tuesday February 10th. Please start answering each question on a new page of the answer booklet. EE376A - Information Theory Midterm, Tuesday February 10th Instructions: You have two hours, 7PM - 9PM The exam has 3 questions, totaling 100 points. Please start answering each question on a new page

More information

Lossy Distributed Source Coding

Lossy Distributed Source Coding Lossy Distributed Source Coding John MacLaren Walsh, Ph.D. Multiterminal Information Theory, Spring Quarter, 202 Lossy Distributed Source Coding Problem X X 2 S {,...,2 R } S 2 {,...,2 R2 } Ẑ Ẑ 2 E d(z,n,

More information

X 1 : X Table 1: Y = X X 2

X 1 : X Table 1: Y = X X 2 ECE 534: Elements of Information Theory, Fall 200 Homework 3 Solutions (ALL DUE to Kenneth S. Palacio Baus) December, 200. Problem 5.20. Multiple access (a) Find the capacity region for the multiple-access

More information

An introduction to basic information theory. Hampus Wessman

An introduction to basic information theory. Hampus Wessman An introduction to basic information theory Hampus Wessman Abstract We give a short and simple introduction to basic information theory, by stripping away all the non-essentials. Theoretical bounds on

More information

Multivariate distributions

Multivariate distributions CHAPTER Multivariate distributions.. Introduction We want to discuss collections of random variables (X, X,..., X n ), which are known as random vectors. In the discrete case, we can define the density

More information

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS PROBABILITY AND INFORMATION THEORY Dr. Gjergji Kasneci Introduction to Information Retrieval WS 2012-13 1 Outline Intro Basics of probability and information theory Probability space Rules of probability

More information

Introduction to Probability and Statistics (Continued)

Introduction to Probability and Statistics (Continued) Introduction to Probability and Statistics (Continued) Prof. Nicholas Zabaras Center for Informatics and Computational Science https://cics.nd.edu/ University of Notre Dame Notre Dame, Indiana, USA Email:

More information

Bivariate distributions

Bivariate distributions Bivariate distributions 3 th October 017 lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.) Bivariate Distributions of the Discrete Type The Correlation Coefficient

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

EE5319R: Problem Set 3 Assigned: 24/08/16, Due: 31/08/16

EE5319R: Problem Set 3 Assigned: 24/08/16, Due: 31/08/16 EE539R: Problem Set 3 Assigned: 24/08/6, Due: 3/08/6. Cover and Thomas: Problem 2.30 (Maimum Entropy): Solution: We are required to maimize H(P X ) over all distributions P X on the non-negative integers

More information

Solutions to Homework Set #5 (Prepared by Lele Wang) MSE = E [ (sgn(x) g(y)) 2],, where f X (x) = 1 2 2π e. e (x y)2 2 dx 2π

Solutions to Homework Set #5 (Prepared by Lele Wang) MSE = E [ (sgn(x) g(y)) 2],, where f X (x) = 1 2 2π e. e (x y)2 2 dx 2π Solutions to Homework Set #5 (Prepared by Lele Wang). Neural net. Let Y X + Z, where the signal X U[,] and noise Z N(,) are independent. (a) Find the function g(y) that minimizes MSE E [ (sgn(x) g(y))

More information