ECE 534: Elements of Information Theory. Solutions to Midterm Exam (Spring 2006)

Similar documents
EGR 544 Communication Theory

Chapter 7 Channel Capacity and Coding

Chapter 7 Channel Capacity and Coding

THE ARIMOTO-BLAHUT ALGORITHM FOR COMPUTATION OF CHANNEL CAPACITY. William A. Pearlman. References: S. Arimoto - IEEE Trans. Inform. Thy., Jan.

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Error Probability for M Signals

Lecture 14 (03/27/18). Channels. Decoding. Preview of the Capacity Theorem.

Lecture 3: Shannon s Theorem

Entropy Coding. A complete entropy codec, which is an encoder/decoder. pair, consists of the process of encoding or

Applied Stochastic Processes

Lossy Compression. Compromise accuracy of reconstruction for increased compression.

State Amplification and State Masking for the Binary Energy Harvesting Channel

6. Stochastic processes (2)

6. Stochastic processes (2)

Channel Encoder. Channel. Figure 7.1: Communication system

Homework Assignment 3 Due in class, Thursday October 15

ECE559VV Project Report

Simultaneous Optimization of Berth Allocation, Quay Crane Assignment and Quay Crane Scheduling Problems in Container Terminals

Problem Set 9 Solutions

Learning Theory: Lecture Notes

Probability Theory. The nth coefficient of the Taylor series of f(k), expanded around k = 0, gives the nth moment of x as ( ik) n n!

Stanford University CS254: Computational Complexity Notes 7 Luca Trevisan January 29, Notes for Lecture 7

Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING

Mathematical Models for Information Sources A Logarithmic i Measure of Information

Errors for Linear Systems

Introduction to Information Theory, Data Compression,

Lecture 4: Universal Hash Functions/Streaming Cont d

Introduction to information theory and data compression

Numerical Heat and Mass Transfer

Calculation of time complexity (3%)

The Minimum Universal Cost Flow in an Infeasible Flow Network

xp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ

S Advanced Digital Communication (4 cr) Targets today

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 12 10/21/2013. Martingale Concentration Inequalities and Applications

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

Compression in the Real World :Algorithms in the Real World. Compression in the Real World. Compression Outline

Quantum and Classical Information Theory with Disentropy

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

Lec 02 Entropy and Lossless Coding I

CSCE 790S Background Results

Lecture 10: May 6, 2013

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

EM and Structure Learning

Chapter 8 SCALAR QUANTIZATION

% & 5.3 PRACTICAL APPLICATIONS. Given system, (49) , determine the Boolean Function, , in such a way that we always have expression: " Y1 = Y2

Lecture 5 Decoding Binary BCH Codes

Assortment Optimization under MNL

The Two-User Gaussian Fading Broadcast Channel

3.1 ML and Empirical Distribution

Continuous Time Markov Chain

14 Lagrange Multipliers

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

Expected Value and Variance

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017

Limited Dependent Variables

Rao IIT Academy/ SSC - Board Exam 2018 / Mathematics Code-A / QP + Solutions JEE MEDICAL-UG BOARDS KVPY NTSE OLYMPIADS SSC - BOARD

Markov chains. Definition of a CTMC: [2, page 381] is a continuous time, discrete value random process such that for an infinitesimal

Winter 2008 CS567 Stochastic Linear/Integer Programming Guest Lecturer: Xu, Huan

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

ON A DETERMINATION OF THE INITIAL FUNCTIONS FROM THE OBSERVED VALUES OF THE BOUNDARY FUNCTIONS FOR THE SECOND-ORDER HYPERBOLIC EQUATION

Introduction to Random Variables

Lecture Notes on Linear Regression

THE ROYAL STATISTICAL SOCIETY 2006 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE

Lecture Space-Bounded Derandomization

Finding Dense Subgraphs in G(n, 1/2)

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14

763622S ADVANCED QUANTUM MECHANICS Solution Set 1 Spring c n a n. c n 2 = 1.

1 The Mistake Bound Model

Digital Signal Processing

Marginal Effects in Probit Models: Interpretation and Testing. 1. Interpreting Probit Coefficients

8.592J: Solutions for Assignment 7 Spring 2005

SELECTED SOLUTIONS, SECTION (Weak duality) Prove that the primal and dual values p and d defined by equations (4.3.2) and (4.3.3) satisfy p d.

Some modelling aspects for the Matlab implementation of MMA

Hidden Markov Models

Assignment 2. Tyler Shendruk February 19, 2010

HMMT February 2016 February 20, 2016

High resolution entropy stable scheme for shallow water equations

A random variable is a function which associates a real number to each element of the sample space

Markov Chain Monte Carlo Lecture 6

Module 9. Lecture 6. Duality in Assignment Problems

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family

Complete subgraphs in multipartite graphs

The exam is closed book, closed notes except your one-page cheat sheet.

The Feynman path integral

Engineering Risk Benefit Analysis

Entropy of Markov Information Sources and Capacity of Discrete Input Constrained Channels (from Immink, Coding Techniques for Digital Recorders)

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD

A new construction of 3-separable matrices via an improved decoding of Macula s construction

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

Tokyo Institute of Technology Periodic Sequencing Control over Multi Communication Channels with Packet Losses

Matrix Approximation via Sampling, Subspace Embedding. 1 Solving Linear Systems Using SVD

Tornado and Luby Transform Codes. Ashish Khisti Presentation October 22, 2003

Problem Set 9 - Solutions Due: April 27, 2005

Math 426: Probability MWF 1pm, Gasson 310 Homework 4 Selected Solutions

Communication Complexity 16:198: February Lecture 4. x ij y ij

College of Computer & Information Science Fall 2009 Northeastern University 20 October 2009

6.3.4 Modified Euler s method of integration

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Transcription:

ECE 534: Elements of Informaton Theory Solutons to Mdterm Eam (Sprng 6) Problem [ pts.] A dscrete memoryless source has an alphabet of three letters,, =,, 3, wth probabltes.4,.4, and., respectvely. (a) Fnd the bnary Huffman code for ths source and determne the average number of bts needed for each source letter. (b) Suppose two letters at a tme are encoded nto a bnary sequence. Fnd the Huffman code and the average number of bts needed per source letter. Soluton: (a) One possble Huffman code s: C() =, C() =, and C(3) =. The average number of bts per source letter s.4 + (.4 +.) =.6. (b) One possble code constructon s: C() =, C() =, C(3) =, C() =, C() =, C(3) =, C(3) =, C(3) =, and C(33) =. The average number of bts per source letter s [3 + (.8 +.4)]/ =.56. Problem [ pts.] A source X produces letters from a three-symbol alphabet wth the probablty assgnment P X () = /4, P X () = /4, and P X () = /. Each source letter s transmtted through two channels smultaneously wth outputs y and z and the transton probabltes ndcated below: P(y ) y P(z ) z Calculate H(X), H(Y ), H(Z), H(Y,Z), I(X;Y ), I(X;Z), I(X;Y Z), and I(X;Y,Z). Soluton: H(X) = 4 log 4 + 4 log 4 + log =.5

The probablty dstrbuton of Y s P Y () = + =. Therefore, 4 H(Y ) = Smlarly, P Z () = P Z () = and H(Z) = If Z =, then X =, and Y =, wth equal probablty; f Z =, then X =, wth equal probablty and as a consequence, Y =, wth equal probablty. Therefore, H(Y Z) = P Z ()H(Y Z = ) + P Z ()H(Y Z = ) = and H(Y,Z) = H(Z) + H(Y Z) = + = Snce H(Y X = ) = H(Y X = ) = and H(Y X = ) =, we have I(X;Y ) = H(Y ) H(Y X) = =.5 Snce H(Z X) =, we have I(X;Z) = H(Z) H(Z X) = H(Z) = Snce Z s completely determned by X, H(Y X,Z) = H(Y X) =.5. We have I(X;Y Z) = H(Y Z) H(Y X,Z) =.5 =.5 and I(X;Y,Z) = I(X;Z) + I(X;Y Z) = +.5 =.5 Problem 3 [3 pts.] a) Prove that the number of elements n the δ-typcal set A(n,δ) satsfes A(n,δ) nh(x)+nδ. b) Prove that A(n,δ) ( δ) nh(x) nδ for suffcently large n. c) Prove that the epected length L of a D-ary pref code for a random varable X satsfes L log D H(X). (Hnt: use Kraft s Inequalty D l.) Soluton: a) A(n,δ) contans the set of sequences (,..., n ), such that n n = log P X ( H(X) δ. Therefore, ) every sequence n A(n, δ) satsfes: Usng the lower bound n (), we have and hence P[A(n,δ)] = nh(x) nδ P(,..., n ) nh(x)+nδ () (,..., n) A(n,δ) P(,..., n ) A(n,δ) nh(x) nδ A(n,δ) nh(x)+nδ

3 b) We have from the law of large numbers that P[A(n,δ)] δ for all suffcently large n. Usng the upper bound n (), we have for all suffcently large n, δ P[A(n,δ)] = P(,..., n ) A(n,δ) nh(x)+nδ and hence (,..., n) A(n,δ) A(n,δ) ( δ) nh(x) nδ c) Denotng the length of the codeword for a A as l a, we have L log D H(X) = a A l a log D a A log = log D la a A (log e) ( D l a ) a A ( ) = (log e) D la a where the frst nequalty comes from ln and the last nequalty comes from Kraft s nequalty. Problem 4 [3 pts.] Consder n dscrete memoryless channels wth capactes C,C,...,C n, respectvely. Both the nput and the output alphabet sets of dfferent channels are dsjont. We defne the sum channel of these n channels as a channel that has all n channels avalable for use but only one channel may be used at any gven tme. (a) Prove that the capacty of the sum channel s gven by n C = log C (b) Use the above result to fnd the capacty of the followng channel: = P(y ) ɛ ɛ ɛ ɛ y

4 Soluton: (a) Denote the nput and output of the th channel as X and Y. Denote the nput and output alphabets of the th channel as A and B, respectvely. Denote the nput and output of the sum channel as X and Y. Suppose the th channel s used wth probablty p, and the nput dstrbuton for the th channel s P X. The nput dstrbuton of the sum channel s: { p P X () A, =,...,n P X () =. otherwse The output dstrbuton s therefore { p A P Y (y) = P X ()P Y X (y ) y B, =,...,n otherwse Snce P Y (y) = A P X ()P Y X (y ), we have H(Y ) = y P Y (y) log P Y (y) y B p P Y (y) log p P Y (y). p log p + p H(Y ). The condtonal entropy s H(Y X) = P X ()H(Y X = ) p P X ()H(Y X = ) A p H(Y X ). The mutual nformaton can then be obtaned as I(X;Y ) = H(Y ) H(Y X) p log p + p I(X ;Y ). To mamze the mutual nformaton, we need to choose P X such that the capacty of the th channel s acheved, leadng to The Lagrangan of ths optmzaton problem s C = supi(x;y ) P X [ = sup p log + (p,...,p n) p p C ] L(p,...,p n,λ) p log p + p C + λ( p )

5 Solvng the followng equatons: L = log log e + C + λ =, p p L λ = p =, =,...,n we obtan the optmal values of (p,...,p n ) as p = C C, =,...,n. The channel capacty s thus C = log C (b) Ths channel can be decomposed nto the sum of a bnary symmetrc channel wth capacty h(ɛ) and a channel wth zero capacty. The capacty of the sum channel s thus C = log ( + h(ɛ))