Homework Set #2 Data Compression, Huffman code and AEP

Similar documents
Solutions to Set #2 Data Compression, Huffman code and AEP

EE376A: Homework #2 Solutions Due by 11:59pm Thursday, February 1st, 2018

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018

COMM901 Source Coding and Compression. Quiz 1

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16

Coding of memoryless sources 1/35

Communications Theory and Engineering

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria

= 4. e t/a dt (2) = 4ae t/a. = 4a a = 1 4. (4) + a 2 e +j2πft 2

Lecture 22: Final Review

LECTURE 13. Last time: Lecture outline

4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science Transmission of Information Spring 2006

Lecture 3. Mathematical methods in communication I. REMINDER. A. Convex Set. A set R is a convex set iff, x 1,x 2 R, θ, 0 θ 1, θx 1 + θx 2 R, (1)

Information Theory. M1 Informatique (parcours recherche et innovation) Aline Roumy. January INRIA Rennes 1/ 73

National University of Singapore Department of Electrical & Computer Engineering. Examination for

Chapter 5: Data Compression

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye

SIGNAL COMPRESSION Lecture Shannon-Fano-Elias Codes and Arithmetic Coding

Lecture 16. Error-free variable length schemes (contd.): Shannon-Fano-Elias code, Huffman code

EE376A - Information Theory Midterm, Tuesday February 10th. Please start answering each question on a new page of the answer booklet.

Chapter 3 Source Coding. 3.1 An Introduction to Source Coding 3.2 Optimal Source Codes 3.3 Shannon-Fano Code 3.4 Huffman Code

Exercises with solutions (Set B)

EE376A - Information Theory Final, Monday March 14th 2016 Solutions. Please start answering each question on a new page of the answer booklet.

3F1 Information Theory, Lecture 3

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1

3F1 Information Theory, Lecture 3

10-704: Information Processing and Learning Fall Lecture 9: Sept 28

Chapter 2 Date Compression: Source Coding. 2.1 An Introduction to Source Coding 2.2 Optimal Source Codes 2.3 Huffman Code

10-704: Information Processing and Learning Fall Lecture 10: Oct 3

Capacity of a channel Shannon s second theorem. Information Theory 1/33

Information Theory. David Rosenberg. June 15, New York University. David Rosenberg (New York University) DS-GA 1003 June 15, / 18

Multimedia Communications. Mathematical Preliminaries for Lossless Compression

PART III. Outline. Codes and Cryptography. Sources. Optimal Codes (I) Jorge L. Villar. MAMME, Fall 2015

Information Theory. Week 4 Compressing streams. Iain Murray,

Motivation for Arithmetic Coding

Information Theory. Lecture 5 Entropy rate and Markov sources STEFAN HÖST

Lecture 5: Asymptotic Equipartition Property

Data Compression. Limit of Information Compression. October, Examples of codes 1

ELEC546 Review of Information Theory

Information Theory and Statistics Lecture 2: Source coding

Lecture 6: Kraft-McMillan Inequality and Huffman Coding

Communication Theory and Engineering

Lecture 2: August 31

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions

Lecture 3: Channel Capacity

Chapter 2: Source coding

EE5585 Data Compression May 2, Lecture 27

Entropy as a measure of surprise

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

Homework 1 Due: Thursday 2/5/2015. Instructions: Turn in your homework in class on Thursday 2/5/2015

Lecture 3 : Algorithms for source coding. September 30, 2016

Lecture 1: Introduction, Entropy and ML estimation

EE5319R: Problem Set 3 Assigned: 24/08/16, Due: 31/08/16

lossless, optimal compressor

EECS 229A Spring 2007 * * (a) By stationarity and the chain rule for entropy, we have

EE 376A: Information Theory Lecture Notes. Prof. Tsachy Weissman TA: Idoia Ochoa, Kedar Tatwawadi

COMPSCI 650 Applied Information Theory Jan 21, Lecture 2

Information measures in simple coding problems

X 1 : X Table 1: Y = X X 2

1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H.

LECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem

MARKOV CHAINS A finite state Markov chain is a sequence of discrete cv s from a finite alphabet where is a pmf on and for

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

Lecture 8: Channel and source-channel coding theorems; BEC & linear codes. 1 Intuitive justification for upper bound on channel capacity

Lecture 11: Quantum Information III - Source Coding

Information Theory: Entropy, Markov Chains, and Huffman Coding

MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016

10-704: Information Processing and Learning Spring Lecture 8: Feb 5

Chapter 9 Fundamental Limits in Information Theory

Chaos, Complexity, and Inference (36-462)

Quiz 2 Date: Monday, November 21, 2016

Summary of Last Lectures

Shannon s noisy-channel theorem

CSCI 2570 Introduction to Nanocomputing

EE376A: Homeworks #4 Solutions Due on Thursday, February 22, 2018 Please submit on Gradescope. Start every question on a new page.

Intro to Information Theory

EE5585 Data Compression January 29, Lecture 3. x X x X. 2 l(x) 1 (1)

ECS 452: Digital Communication Systems 2015/2. HW 1 Due: Feb 5

Electrical and Information Technology. Information Theory. Problems and Solutions. Contents. Problems... 1 Solutions...7

Solutions to Homework Set #1 Sanov s Theorem, Rate distortion

Lecture 4 Channel Coding

INFORMATION THEORY AND STATISTICS

Entropy, Relative Entropy and Mutual Information Exercises

Lecture Notes on Digital Transmission Source and Channel Coding. José Manuel Bioucas Dias

LECTURE 10. Last time: Lecture outline

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

The Communication Complexity of Correlation. Prahladh Harsha Rahul Jain David McAllester Jaikumar Radhakrishnan

Basic Principles of Lossless Coding. Universal Lossless coding. Lempel-Ziv Coding. 2. Exploit dependences between successive symbols.

Lecture 1. Introduction

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Lecture 14 February 28

ECE Advanced Communication Theory, Spring 2009 Homework #1 (INCOMPLETE)

Entropies & Information Theory

Lecture 4 Noisy Channel Coding

Lecture 1 : Data Compression and Entropy

Run-length & Entropy Coding. Redundancy Removal. Sampling. Quantization. Perform inverse operations at the receiver EEE

ECE 587 / STA 563: Lecture 5 Lossless Compression

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016

1 Introduction to information theory

Transcription:

Homework Set #2 Data Compression, Huffman code and AEP 1. Huffman coding. Consider the random variable ( x1 x X = 2 x 3 x 4 x 5 x 6 x 7 0.50 0.26 0.11 0.04 0.04 0.03 0.02 (a Find a binary Huffman code for X. (b Find the expected codelength for this encoding. (c Extend the Binary Huffman method to Ternarry (Alphabet of 3 and apply it for X. 2. Codes. Let X 1, X 2,..., i.i.d. with 1, with probability 1/2 X = 2, with probability 1/4 3, with probability 1/4. Consider the code assignment 0, if x = 1 C(x = 01, if x = 2 11, if x = 3. (a Is this code nonsingular? (b Uniquely decodable? (c Instantaneous? (d Entropy Rate is defined as H(X n H(X lim. (1 n n What is the entropy rate of the process Z 1 Z 2 Z 3... = C(X 1 C(X 2 C(X 3...? 3. Huffman via Matlab 1

(a Give a Huffman encoding into an alphabet of size D = 2 of the following probability mass function: ( 1 2, 1 8, 1 8, 1 16, 1 16, 1 16, 1 16 (b Assume you have a file of size 1,000 symbols where the symbols are distributed i.i.d. according to the pmf above. After applying the Huffman code, what would be the pmf of the compressed binary file (namely, what is the probability of 0 and 1 in the compressed file, and what would be the expected length? (c Generate a sequence (using Matlab or any other software of 10,000 symbols of X with i.i.d. probability P X. Assume the alphabet of X is X = (0, 1,..., 6. (d What is the percentage of each symbol (0, 1,..., 6 in the sequence that was generated by Matlab. Explain this result using the law of large numbers. (e Represent each symbol in X using a simple binary representation. Namely, X = 0 represent as 000, X = 1 represent as 001,X = 2 represent as 010,..., X = 6 represent as 110. (f What is the length of the simple representation. What percentage of 0 and 1 do you have in this representation? (g Now, compress the 10,000 symbols of X, into bits using Huffman code. (h What is the length of the compressed file. What percentage of 0 and 1 do you have in this representation? (i Explain the results. Note: You may use the MATLAB functions randsample and strfind. 4. Entropy and source coding of a source with infinite alphabet Let X be an i.i.d. random variable with an infinite alphabet, X = {1, 2, 3,...}. In addition let P (X = i = 2 i. (a What is the entropy of the random variable? (b Find an optimal variable length code, and show that it is indeed optimal. 5. Bad wine. One is given 6 bottles of wine. It is known that precisely one bottle has gone bad (tastes terrible. From inspection of the bottles it is determined that the probability p i that the i th bottle is bad is given by (p 1, p 2,..., p 6 = ( 7, 5, 4, 4, 3, 3. Tasting will determine the 26 26 26 26 26 26 bad wine. Suppose you taste the wines one at a time. Choose the order of tasting to minimize the expected number of tastings required to determine the bad bottle. Remember, if the first 5 wines pass the test you don t have to taste the last. 2

(a What is the expected number of tastings required? (b Which bottle should be tasted first? Now you get smart. For the first sample, you mix some of the wines in a fresh glass and sample the mixture. You proceed, mixing and tasting, stopping when the bad bottle has been determined. (c What is the minimum expected number of tastings required to determine the bad wine? (d What mixture should be tasted first? 6. Relative entropy is cost of miscoding. Let the random variable X have five possible outcomes {1, 2, 3, 4, 5}. Consider two distributions on this random variable Symbol p(x q(x C 1 (x C 2 (x 1 1/2 1/2 0 0 2 1/4 1/8 10 100 3 1/8 1/8 110 101 4 1/16 1/8 1110 110 5 1/16 1/8 1111 111 (a Calculate H(p, H(q, D(p q and D(q p. (b The last two columns above represent codes for the random variable. Verify that the average length of C 1 under p is equal to the entropy H(p. Thus C 1 is optimal for p. Verify that C 2 is optimal for q. (c Now assume that we use code C 2 when the distribution is p. What is the average length of the codewords. By how much does it exceed the entropy H(p? (d What is the loss if we use code C 1 when the distribution is q? 7. Shannon code. Consider the following method for generating a code for a random variable X which takes on m values {1, 2,..., m} with probabilities p 1, p 2,..., p m. Assume that the probabilities are ordered so that p 1 p 2 p m. Define i 1 F i = p i, (2 k=1 the sum of the probabilities of all symbols less than i. Then the codeword for i is the number F i [0, 1] rounded off to l i bits, where l i = log 1 p i. 3

(a Show that the code constructed by this process is prefix-free and the average length satisfies H(X L < H(X + 1. (3 (b Construct the code for the probability distribution (0.5, 0.25, 0.125, 0.125. 8. An AEP-like limit. Let X 1, X 2,... be i.i.d. drawn according to probability mass function p(x. Find lim n [p(x 1, X 2,..., X n ] 1 n. 9. AEP. Let X 1, X 2,... be independent identically distributed random variables drawn according to the probability mass function p(x, x {1, 2,..., m}. Thus p(x 1, x 2,..., x n = n i=1 p(x i. We know that 1 n log p(x 1, X 2,..., X n H(X in probability. Let q(x 1, x 2,..., x n = n i=1 q(x i, where q is another probability mass function on {1, 2,..., m}. (a Evaluate lim 1 n log q(x 1, X 2,..., X n, where X 1, X 2,... are i.i.d. p(x. (b Now evaluate the limit of the log likelihood ratio 1 n log q(x 1,...,X n p(x 1,...,X n when X 1, X 2,... are i.i.d. p(x. Thus the odds favouring q are exponentially small when p is true. 10. drawing a codebook Let X i be a r.v. i.i.d distributed according to P (x. We draw codebook of 2 nr codewords of X n independently using P (x and i.i.d.. We would like to answer the question: what is the probability that the first codeword would be identical to another codeword in the codebook as n goes to infinity. Let x n be a sequence in the typical set A n ɛ (X. What is the asymptotic probability (you may provide an upper and lower bound as n that we draw a sequence X n i.i.d distributed according to P (x and we get x n. Using your answer from the previous sub-question find an α such that if R < α the probability that the first codewaord in the codebook appears twice or more in the codebook goes to zero as n. Find an α such that if R > α the probability that the first codewaord in the codebook appears twice or more in the codebook goes to 1 as n. (Hint: you may use Bernoulli s inequality (1 + x r e rx for all real numbers r 0, x 1 11. Saving the princess A princess was abducted and was put in one of K rooms. Each room is labeled by a number 1,2,...,K. Each room is of a size s i where i = 1, 2,..., k. The probability of the princess to be in room i, p i, is proportional to the size of the room, namely p i = αs i where α is a constant. 4

(a Find α (b In order to save the princess you need to find in which room she is. You may ask the demon an yes/no question. Like is she in room number 1 or is she in room 2 or 5 or is she in a room of odd number, and so on. You will save the princess if only if the expected number of questions is the minimum possible. What would be the questions you should ask the demon to save the princess? 12. Lossless source coding with side information. Consider the lossless source coding with side information that is available at the encoder and decoder, where the source X and the side information Y are i.i.d. P X,Y (x, y. X Encoder M Decoder H Y Y Figure 1: Lossless source coding with side information at the encoder and decoder. Show that a code with rate R < H(X Y can not be achievable, and interpret the result. Hint: Let T f(x n, Y n. Consider nr H(T H(T Y n, (4 and use similar steps, including Fano s inequality, as we used in the class to prove the converse where side information was not available. 13. Challenge: Optimal code for an infinite alphabet This is a bonus question! Solving this question would add 10 points to the final grade of the course. Let X be a r.v. with a discrete and infinite alphabet, in particular assume that X = 1, 2, 3,... The pmf is given by an infinite vector [p 1, p 2, p 3,...] where p i p j if i > j. Find an optimal prefix code for X. 14. Conditional Information Divergence 5

(a Let X, Z be random variables jointly distributed according to P X,Z. We define the conditional informational divergence as follows: D ( ( P X Z PX Z QX Z PZ (x z = P X,Z (x, z log. Q X Z (x z (x,z X Z With respect to this definition, prove for each relation if it is true or false: For any pair of random variables A, B that are jointly distributed according to P A,B, i. ii. iii. iv. (b Consider the setting in Fig. 2. D ( P A,B QA,B = D ( PA QA + D ( PB A QB A PA. D ( P A,B PA P B = D ( PB A PB PA. I(A; B = D ( P B A PB PA. D ( P A B Q A B P B = P B (bd ( P A B=b Q A B=b. b B B B A E F C D Figure 2: Source coding with side information. We would like to compress the source sequence X n losslessly using a prefix code with side information Z n which is available to the encoder and the decoder. The sources (X n, Z n are distributed i.i.d. according to P X,Z and that all the distribution and conditional distributions are dyadic (i.e., P X is dyadic if P X (x = 2 i, for some i, for all x X. We denote the average number of bits per symbol needed to compress the source X n as L. i. What is the minimal L? ii. Although the distribution of (X n, Z n is P X,Z, the distribution that is used design the optimal prefix code is Q X Z P Z. What is the actual L (average bits per symbol of this code? 6

iii. Now, the distribution that is used to design the prefix code is Q X,Z. What is the actual L now? 15. True or False of a constrained inequality: Given are three discrete random variables X, Y, Z that satisfy H(Y X, Z = 0. (a Copy the next relation and write true or false (If true, prove the statement, and if not provide a counterexample. I(X; Y H(Y H(Z (b What are the conditions for which the equality I(X; Y = H(Y H(Z holds. (c Assume that the conditions for I(X; Y = H(Y H(Z are satisfied. Is it true that there exists a function such that Z = g(y? 16. True or False: Copy each relation and write true or false. (a Let X Y Z W be a Markov chain, then the following holds: I(X; W I(Y ; Z. (b For two probability distributions, p XY and q XY, that are defined on X Y, the following holds: D(p XY q XY D(p X q X. (c If X and Y are dependent and also Y and Z are dependent, then X and Z are dependent. 17. Huffman Code : Let X n be a an i.i.d. source that is distributed according to p X : (a Find H(X. x 0 1 2 3 p X (x 0.5 0.25 0.125 0.125 (b Build a binary Huffman code for the source X. (c What is the expected length of the resulting compressed sequence. (d What is the expected number of zeros in the resulting compressed sequence. (e Let X n be an another source distributed i.i.d. according to p X. x 0 1 2 3 p X( x 0.3 0.4 0.1 0.2 What is the expected length of compressing the source X using the code constructed in (b. 7

(f Answer (d for the code constructed in (b and the source X n. (g Is the relative portion of zeros (the quantity in (d divided by the quantity in (c after compressing the source X n and the source X n different? For both sources, explain why there is or there is not a difference. 8