Network coding for multicast relation to compression and generalization of Slepian-Wolf

Similar documents
LECTURE 13. Last time: Lecture outline

EE5585 Data Compression May 2, Lecture 27

Lecture 3: Channel Capacity

Capacity of a channel Shannon s second theorem. Information Theory 1/33

LECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem

The Capacity Region for Multi-source Multi-sink Network Coding

Shannon s noisy-channel theorem

Lecture 15: Conditional and Joint Typicaility

LECTURE 10. Last time: Lecture outline

EE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Distributed Source Coding Using LDPC Codes

ELEC546 Review of Information Theory

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

A Graph-based Framework for Transmission of Correlated Sources over Multiple Access Channels

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions

(Classical) Information Theory III: Noisy channel coding

Shannon s Noisy-Channel Coding Theorem

ECE Information theory Final (Fall 2008)

Lecture 8: Channel and source-channel coding theorems; BEC & linear codes. 1 Intuitive justification for upper bound on channel capacity

Distributed Lossless Compression. Distributed lossless compression system

Information Theory. Lecture 10. Network Information Theory (CT15); a focus on channel capacity results

EE376A: Homeworks #4 Solutions Due on Thursday, February 22, 2018 Please submit on Gradescope. Start every question on a new page.

Shannon s Noisy-Channel Coding Theorem

Lecture 4 Noisy Channel Coding

MARKOV CHAINS A finite state Markov chain is a sequence of discrete cv s from a finite alphabet where is a pmf on and for

Distributed Functional Compression through Graph Coloring

SOURCE CODING WITH SIDE INFORMATION AT THE DECODER (WYNER-ZIV CODING) FEB 13, 2003

Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets

The Communication Complexity of Correlation. Prahladh Harsha Rahul Jain David McAllester Jaikumar Radhakrishnan

Frans M.J. Willems. Authentication Based on Secret-Key Generation. Frans M.J. Willems. (joint work w. Tanya Ignatenko)

Solutions to Homework Set #1 Sanov s Theorem, Rate distortion

4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information

Lecture 11: Quantum Information III - Source Coding

Universal Incremental Slepian-Wolf Coding

X 1 : X Table 1: Y = X X 2

On Multiple User Channels with State Information at the Transmitters

Interactive Communication for Data Exchange

Source and Channel Coding for Correlated Sources Over Multiuser Channels

EE 4TM4: Digital Communications II. Channel Capacity

The Method of Types and Its Application to Information Hiding

ECE Information theory Final

Lecture 22: Final Review

Lecture 11: Continuous-valued signals and differential entropy

Lecture Notes for Communication Theory

Communications Theory and Engineering

Lecture 5: Channel Capacity. Copyright G. Caire (Sample Lectures) 122

National University of Singapore Department of Electrical & Computer Engineering. Examination for

(Classical) Information Theory II: Source coding

Representation of Correlated Sources into Graphs for Transmission over Broadcast Channels

Exchanging Third-Party Information in a Network

On Large Deviation Analysis of Sampling from Typical Sets

Slepian-Wolf Code Design via Source-Channel Correspondence

EECS 750. Hypothesis Testing with Communication Constraints

Quantum Information Chapter 10. Quantum Shannon Theory

Hypothesis Testing with Communication Constraints

Cases Where Finding the Minimum Entropy Coloring of a Characteristic Graph is a Polynomial Time Problem

On Capacity Under Received-Signal Constraints

Network Coding on Directed Acyclic Graphs

On Function Computation with Privacy and Secrecy Constraints

Noisy channel communication

Source Coding and Function Computation: Optimal Rate in Zero-Error and Vanishing Zero-Error Regime

Lecture 11: Polar codes construction

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

Lecture 1: Introduction, Entropy and ML estimation

Reliable Computation over Multiple-Access Channels

A Practical and Optimal Symmetric Slepian-Wolf Compression Strategy Using Syndrome Formers and Inverse Syndrome Formers

Solutions to Homework Set #2 Broadcast channel, degraded message set, Csiszar Sum Equality

Lossy Distributed Source Coding

Interactive Decoding of a Broadcast Message

Coding for Computing. ASPITRG, Drexel University. Jie Ren 2012/11/14

Electrical and Information Technology. Information Theory. Problems and Solutions. Contents. Problems... 1 Solutions...7

On Scalable Coding in the Presence of Decoder Side Information

A Novel Asynchronous Communication Paradigm: Detection, Isolation, and Coding

Exercise 1. = P(y a 1)P(a 1 )

Information Theory and Statistics, Part I

Homework Set #2 Data Compression, Huffman code and AEP

Lecture 5: Asymptotic Equipartition Property

for some error exponent E( R) as a function R,

Multiterminal Source Coding with an Entropy-Based Distortion Measure

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018

Entropies & Information Theory

Lecture 2. Capacity of the Gaussian channel

16.36 Communication Systems Engineering

Note that the new channel is noisier than the original two : and H(A I +A2-2A1A2) > H(A2) (why?). min(c,, C2 ) = min(1 - H(a t ), 1 - H(A 2 )).

Graph Coloring and Conditional Graph Entropy

9 Forward-backward algorithm, sum-product on factor graphs

Notes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel

Lecture 14 February 28

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

Information Theory: Entropy, Markov Chains, and Huffman Coding

Secret Key and Private Key Constructions for Simple Multiterminal Source Models

A Comparison of Two Achievable Rate Regions for the Interference Channel

Second-Order Asymptotics in Information Theory

Information Theory and Coding Techniques: Chapter 1.1. What is Information Theory? Why you should take this course?

An introduction to basic information theory. Hampus Wessman

EE 376A: Information Theory Lecture Notes. Prof. Tsachy Weissman TA: Idoia Ochoa, Kedar Tatwawadi

Common Information. Abbas El Gamal. Stanford University. Viterbi Lecture, USC, April 2014

Shannon s A Mathematical Theory of Communication

Transcription:

Network coding for multicast relation to compression and generalization of Slepian-Wolf 1

Overview Review of Slepian-Wolf Distributed network compression Error exponents Source-channel separation issues Code construction for finite field multiple access networks 2

Distributed data compression Consider two correlated sources (X, Y ) p(x, y) that must be separately encoded for a user who wants to reconstruct both What information transmission rates from each source allow decoding with arbitrarily small probability of error? E.g. H(X 1 ) X 1. X 2. H(X2 X1 ) 3

Distributed source code A ((2 nr 1, 2 nr 2),n) distributed source code for joint source (X, Y ) consists of encoder maps and a decoder map f 1 : X n {1, 2,...,2 nr 1} n f 2 : Y {1, 2,...,2 nr 2} n g : {1, 2,...,2 nr 1} {1, 2,...,2 nr 2} X n Y - X n is mapped to f 1 (X n ) - Y n is mapped to f 2 (Y n ) -(R 1,R 2 ) is the rate pair of the code 4

Probability of error P e (n) =Pr{g(f1 (X n ),f 2 (Y n )) = ( X n,y n )}

Slepian-Wolf Definitions: A rate pair (R 1,R 2 ) is achievable if there exists a sequence of ((2 nr 1, 2 nr 2),n) distributed source codes with probability of error (n) 0as n P e achievable rate region - closure of the set of achievable rates Slepian-Wolf Theorem: 5

For the distributed source coding problem for source (X, Y ) drawn i.i.d. p(x, y), the achievable rate region is R 1 H(X Y ) R 2 H(Y X) R 1 + R 2 H(X, Y )

Proof of achievability Main idea: show that if the rate pair is in the Slepian-Wolf region, we can use a random binning encoding scheme with typical set decoding to obtain a probability of error that tends to zero Coding scheme: Source X assigns every sourceword x X n randomly among 2 nr 1 bins, and source Y independently assigns every y Y n randomly among 2 nr 2 bins Each sends the bin index corresponding to the message 6

the receiver decodes correctly if there is exactly one jointly typical sourceword pair corresponding to the received bin indexes, otherwise it declares an error

Random binning for single source compression An encoder that knows the typical set can compress a source X to H(X)+ɛ without loss, by employing separate codes for typical and atypical sequences Random binning is a way to compress a source X to H(X) + ɛ with asymptotically small probability of error without the encoder knowing the typical set, as well as the decoder knows the typical set the encoder maps each source sequence X n uniformly at random into one of 2 nr bins 7

the bin index, which is R bits long, forms the code the receiver decodes correctly if there is exactly one typical sequence corresponding to the received bin index

Error analysis An error occurs if: a) the transmitted sourceword is not typical, i.e. event (n) E 0 = {X / A ɛ } b) there exists another typical sourceword in the same bin, i.e.event E 1 = { x )= f(x), x (n) = X : f(x A ɛ } Use union of events bound: (n) P e = Pr(E 0 E 1 ) Pr(E 0 )+Pr(E 1 ) 8

Error analysis continued Pr(E 0 ) 0 by the Asymptotic Equipartition Property (AEP) Pr(E 1 ) = Pr{ x = x : f(x )= f(x), x x x x x (n) A ɛ (n) } (n) = A ɛ 2 nr x 2 nr 2 n(h(x)+ɛ) 0if R>H(X) x A ɛ Pr(f(x )= f(x)) 9

For sufficiently large n, Pr(E 0 ), Pr(E 1 ) < ɛ (n) P ɛ < 2ɛ

Jointly typical sequences (n) The set A ɛ of jointly typical sequences is the set of sequences (x, y) X n Y n with probability: 2 n(h(x)+ɛ) p X (x) 2 n(h(x) ɛ) 2 n(h(y )+ɛ) p Y (y) 2 n(h(y ) ɛ) 2 n(h(x,y )+ɛ) p X,Y (x, y) 2 n(h(x,y ) ɛ) for (X, Y) sequences of length n IID according to p X,Y (x, y) = i=1 p X,Y (x i,y i ) n Size of typical set: (n) 2 n(h(x,y )+ɛ) A ɛ 10

Proof: 1 = p(x, y) p(x, y) (n) A ɛ A ɛ (n) 2 n(h(x,y )+ɛ)

Conditionally typical sequences (n) The conditionally typical set A ɛ (X y) for a given typical y sequence is the set of x sequences that are jointly typical with the given y sequence. Size of conditionally typical set: (n) A ɛ (X y) 2 n(h(x Y )+ɛ) Proof: 11

For (x, y) A ( ɛ n) (X, Y ),. p(y) = 2 n(h(y )±ɛ). p(x, y) = 2 n(h(x,y )±ɛ) p(x y) = p(x, y) p(y). = 2 n(h(x Y )±2ɛ) Hence 1 p(x y) (n) x A ɛ (X y) (n) A ɛ 2 n(h(x Y )+2ɛ)

Proof of achievability error analysis Errors occur if: a) the transmitted sourcewords are not jointly typical, i.e. event (n) E 0 = {(X, Y ) / A ɛ } b) there exists another pair of jointly typical sourcewords in the same pair of bins, i.e. one or more of the following events E 1 = { x = X : f 1 (x )= f 1 (X), (x (n), Y) A ɛ } E 2 = { y )= f 2 (Y), (X, y (n) = Y : f2 (y ) A ɛ } E 12 = { (x, y ): x X, y Y,f 1 (x )= f 1 (X), f 2 (y )= f 2 (Y), (x, y ) A ( ɛ n) } 12

Use union of events bound: (n) P e = Pr(E 0 E 1 E 2 E 12 ) Pr(E 0 )+Pr(E 1 )+Pr(E 2 )+Pr(E 12 )

Error analysis continued Pr(E 0 ) 0 bythe AEP Pr(E 1 ) = Pr{ x x : f 1 (x )=f 1 (x), (x,y) (x, y) A (n) = (x,y) (x,y) 2 nr 1 2 x = x (x, y) A (n) ɛ ɛ } A (n) ɛ (X y) 2 nr 1 n(h(x Y )+2ɛ) 0ifR 1 >H(X Y ) Pr(f 1 (x )=f 1 (x)) 13

Similarly, Pr(E 2 ) 2 nr 2 2 n(h(y X)+2ɛ) 0if R 2 >H(Y X) Pr(E 12 ) 2 n(r 1+R 2 ) 2 n(h(x,y )+ɛ) 0if R 1 + R 2 >H(X, Y )

Error analysis continued Thus, if we are in the Slepian-Wolf rate region, for sufficiently large n, Pr(E 0 ), Pr(E 1 ), Pr(E 2 ), Pr(E 12 ) < ɛ (n) P ɛ < 4ɛ Since the average probability of error is less than 4ɛ, there exist at least one code (f 1,f 2,g ) with probability of error < 4ɛ. (n) Thus, there exists a sequence of codes with P ɛ 0. 14

Model for distributed network compression arbitrary directed graph with integer capacity links discrete memoryless source processes with integer bit rates randomized linear network coding over vectors of bits in F 2 coefficients of overall combination transmitted to receivers receivers perform minimum entropy or maximum a posteriori probability decoding 15

Distributed compression problem Consider two sources of bit rates r 1,r 2, whose output values in each unit time period are drawn i.i.d. from the same joint distribution Q linear network coding in F 2 over vectors of nr 1 and nr 2 bits from each source respectively Define 16

m 1 and m 2 the minimum cut capacities between the receiver and each source respectively m 3 the minimum cut capacity between the receiver and both sources L the maximum source-receiver path length

Theorem 1 The error probability at each receiver using minimum entropy or maximum a posteriori probability decoding is at most i 3 =1 p e i,where { ( p 1 e exp n min D(P X1 X 2 Q) X 1,X 2 ) } 1 + + m 1 (1 log L) H(X 1 X 2 ) +2 2r 1+r 2 log(n +1) n { ( p 2 e exp n min D(P X1 X 2 Q) X 1,X 2 ) } 1 + + m 2 (1 log L) H(X 2 X 1 ) +2 r 1+2r 2 log(n +1) n { ( p 3 e exp n min D(P X1 X 2 Q) X 1,X 2 ) } + m 3 (1 1 + log L) H(X 1 X 2 ) +2 2r 1+2r 2 log(n +1) n 17

Distributed compression Redundancy is removed or added in different parts of the network depending on available capacity Achieved without knowledge of source entropy rates at interior network nodes For the special case of a Slepian-Wolf source network consisting of a link from each source to the receiver, the network coding error exponents reduce to known error exponents for linear Slepian-Wolf coding [Csi82] 18

Proof outline Error probability 3 i =1 pi e,where p1 e is the probability of correctly decoding X 2 but not X 1, p2 e is the probability of correctly decoding X 1 but not X 2 p3 e is the probability of wrongly decoding X 1,X 2 Proof approach using method of types similar to that in [Csi82] Types P xi, joint types P xy are the empirical distributions of elements in vectors x i 19

Proof outline (cont d) Bound error probabilities by summing over sets of joint types {P X 1 = X 1, X 2 = X 2 } i = 1 X 1 X 1X 2 X 2 P i n = {P X X 1 = X 1, X 2 X 2 } i = 2 1 X 1X 2 X 2 {P X X 2 = X 2 } 1 X 1X X 2 X 2 1 = X 1, i = 3 nr i where Xi, X i F 2 20

sequences of each type T X1 X 2 = [ x y { } n(r ] F 1 +r 2 ) 2 Pxy = P X1 X 2 { n(r 1 +r 2 ) T X 1X 2 X 1 X 2 (xy) = [ x ỹ ] F 2 P x yxy = P X 1X 2X 1 X 2 }

Proof outline (cont d) Define P i,i =1, 2, the probability that distinct (x, y), ( x, y), where x = x, at the receiver P 3, the probability that (x, y), ( x, y), where x = x,y = y, are mapped to the same output at the receiver These probabilities can be calculated for a given network, or bounded in terms of block length n and network parameters 21

Proof outline (cont d) A link with 1 nonzero incoming signal carries the zero signal with probability 2 1 nc, where c is the link capacity this is equal to the probability that a pair of distinct input values are mapped to the same output on the link We can show by induction on the minimum cut capacities m i that ( P i 1 (1 1 ) mi ) L ( ) L mi 2 n 2 n 22

Proof outline (cont d) We substitute in cardinality bounds P 1 < (n +1) 22r 1 +r 2 n P 2 < (n +1) 2r 1 +2r 2 n P 3 < (n +1) 22r 1 +2r 2 n T X1 X 2 exp{nh(x 1 X 2 )} T X 1X 2 X 1 X (xy) exp{nh(x 1X 2 X 2 1 X 2 )} probability of source vector of type (x, y) T X1 X 2 Q n (xy) = exp{ n(d(p X1 X 2 Q)+ H(X 1 X 2 ))} 23

Proof outline (cont d) and the decoding conditions minimum entropy decoder: H(X 1X 2) H(X 1 X 2 ) maximum a posteriori probability decoder: D(P Q)+ H(X 1X 2) D(P X1 X 2 Q)+ H(X 1 X 2 ) X 1 X 2 to obtain the result 24

Conclusions Distributed randomized network coding can achieve distributed compression of correlated sources Error exponents generalize results for linear Slepian Wolf coding Further work: investigation of non-uniform code distributions, other types of codes, and other decoding schemes 25

MIT OpenCourseWare http://ocw.mit.edu 6.441 Information Theory Spring 2010 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.