Representation of Correlated Sources into Graphs for Transmission over Broadcast Channels

Similar documents
A Graph-based Framework for Transmission of Correlated Sources over Multiple Access Channels

On Large Deviation Analysis of Sampling from Typical Sets

Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels

Computing sum of sources over an arbitrary multiple access channel

Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets

On Multiple User Channels with State Information at the Transmitters

A Comparison of Two Achievable Rate Regions for the Interference Channel

Lattices for Distributed Source Coding: Jointly Gaussian Sources and Reconstruction of a Linear Function

On Scalable Coding in the Presence of Decoder Side Information

Joint Source-Channel Coding for the Multiple-Access Relay Channel

Chain Independence and Common Information

A New Achievable Region for Gaussian Multiple Descriptions Based on Subset Typicality

Shannon s noisy-channel theorem

COMPLEMENTARY graph entropy was introduced

An Achievable Rate Region for the 3-User-Pair Deterministic Interference Channel

Lecture 4 Noisy Channel Coding

Graph Coloring and Conditional Graph Entropy

Simultaneous Nonunique Decoding Is Rate-Optimal

Capacity Region of Reversely Degraded Gaussian MIMO Broadcast Channel

Multiuser Successive Refinement and Multiple Description Coding

Error Exponent Region for Gaussian Broadcast Channels

Cognitive Multiple Access Networks

On Gaussian MIMO Broadcast Channels with Common and Private Messages

New Results on the Equality of Exact and Wyner Common Information Rates

arxiv: v1 [cs.it] 4 Jun 2018

Source and Channel Coding for Correlated Sources Over Multiuser Channels

Network coding for multicast relation to compression and generalization of Slepian-Wolf

Reliable Computation over Multiple-Access Channels

Information Theory. Lecture 10. Network Information Theory (CT15); a focus on channel capacity results

On the Capacity Region for Secure Index Coding

On Scalable Source Coding for Multiple Decoders with Side Information

Degrees of Freedom Region of the Gaussian MIMO Broadcast Channel with Common and Private Messages

Distributed Functional Compression through Graph Coloring

Variable Length Codes for Degraded Broadcast Channels

Amobile satellite communication system, like Motorola s

On The Binary Lossless Many-Help-One Problem with Independently Degraded Helpers

The Capacity of the Semi-Deterministic Cognitive Interference Channel and its Application to Constant Gap Results for the Gaussian Channel

The Poisson Channel with Side Information

5958 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 12, DECEMBER 2010

Simplified Composite Coding for Index Coding

The Capacity Region of a Class of Discrete Degraded Interference Channels

A Half-Duplex Cooperative Scheme with Partial Decode-Forward Relaying

Network Coding on Directed Acyclic Graphs

Secret Key and Private Key Constructions for Simple Multiterminal Source Models

On Lossless Coding With Coded Side Information Daniel Marco, Member, IEEE, and Michelle Effros, Fellow, IEEE

The Capacity Region of the Gaussian Cognitive Radio Channels at High SNR

Lecture 1: The Multiple Access Channel. Copyright G. Caire 12

6196 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 9, SEPTEMBER 2011

An Alternative Proof for the Capacity Region of the Degraded Gaussian MIMO Broadcast Channel

Duality Between Channel Capacity and Rate Distortion With Two-Sided State Information

AN INTRODUCTION TO SECRECY CAPACITY. 1. Overview

A Comparison of Superposition Coding Schemes

On the Capacity Region of the Gaussian Z-channel

Cases Where Finding the Minimum Entropy Coloring of a Characteristic Graph is a Polynomial Time Problem

Secret Key Agreement Using Asymmetry in Channel State Knowledge

II. THE TWO-WAY TWO-RELAY CHANNEL

Index coding with side information

Primary Rate-Splitting Achieves Capacity for the Gaussian Cognitive Interference Channel

The Capacity Region for Multi-source Multi-sink Network Coding

Hypothesis Testing with Communication Constraints

Secret Key Agreement Using Conferencing in State- Dependent Multiple Access Channels with An Eavesdropper

On Compound Channels With Side Information at the Transmitter

An Achievable Error Exponent for the Mismatched Multiple-Access Channel

Distributed Source Coding Using LDPC Codes

On Source-Channel Communication in Networks

IET Commun., 2009, Vol. 3, Iss. 4, pp doi: /iet-com & The Institution of Engineering and Technology 2009

LECTURE 13. Last time: Lecture outline

Distributed Lossy Interactive Function Computation

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information

Performance-based Security for Encoding of Information Signals. FA ( ) Paul Cuff (Princeton University)

Equivalence for Networks with Adversarial State

Multicoding Schemes for Interference Channels

Generalized Writing on Dirty Paper

arxiv: v1 [cs.it] 5 Feb 2016

The Method of Types and Its Application to Information Hiding

On the Secrecy Capacity of Fading Channels

ProblemsWeCanSolveWithaHelper

Decentralized Stochastic Control with Partial Sharing Information Structures: A Common Information Approach

A Novel Asynchronous Communication Paradigm: Detection, Isolation, and Coding

On Common Information and the Encoding of Sources that are Not Successively Refinable

Dispersion of the Gilbert-Elliott Channel

A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources

Optimal Natural Encoding Scheme for Discrete Multiplicative Degraded Broadcast Channels

Upper Bounds on the Capacity of Binary Intermittent Communication

The Gallager Converse

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 62, NO. 5, MAY

Source-Channel Coding Theorems for the Multiple-Access Relay Channel

Arimoto Channel Coding Converse and Rényi Divergence

A Polynomial-Time Algorithm for Pliable Index Coding

Training-Based Schemes are Suboptimal for High Rate Asynchronous Communication

Polar Codes for Some Multi-terminal Communications Problems

The Gaussian Many-Help-One Distributed Source Coding Problem Saurabha Tavildar, Pramod Viswanath, Member, IEEE, and Aaron B. Wagner, Member, IEEE

Variable-Rate Universal Slepian-Wolf Coding with Feedback

Subset Typicality Lemmas and Improved Achievable Regions in Multiterminal Source Coding

Distributed Lossless Compression. Distributed lossless compression system

820 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 58, NO. 2, FEBRUARY Stefano Rini, Daniela Tuninetti, and Natasha Devroye

A Proof of the Converse for the Capacity of Gaussian MIMO Broadcast Channels

5218 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 12, DECEMBER 2006

Remote Source Coding with Two-Sided Information

Interactive Decoding of a Broadcast Message

Transcription:

Representation of Correlated s into Graphs for Transmission over Broadcast s Suhan Choi Department of Electrical Eng. and Computer Science University of Michigan, Ann Arbor, MI 80, USA Email: suhanc@eecs.umich.edu S. Sandeep Pradhan Department of Electrical Eng. and Computer Science University of Michigan, Ann Arbor, MI 80, USA Email: pradhanv@eecs.umich.edu Abstract In this paper we consider the communication problem that involves transmission of correlated sources over broadcast channels. We consider a graph-based framework for this information transmission problem. The system involves a source coding module and a channel coding module. In the source coding module, the sources are efficiently represented using a nearly semi-regular bipartite graph, and the in the channel coding module, the edges of this graph are reliably transmitted over a broadcast channel. We consider nearly semi-regular bipartite graphs as discrete interface between source coding and channel coding in this multiterminal setting. In particular, in this paper, we restrict our attention to the source coding module, building on our earlier work on the channel coding module. We provide an information-theoretic characterization of the rate of growth of the exponent (as a function of the number source samples) of the size of such graphs that can reliably represent a pair of correlated sources. I. INTRODUCTION The broadcast channel, where a transmitter wishes to accomplish reliable simultaneous communication with a set of receivers was first introduced by Cover [] in 97. Since then, the broadcast channel has been studied extensively, and the capacity region for several special classes of broadcast channels have been found. For the discrete memoryless broadcast channel, Marton [] established an inner bound to the capacity region which contains all known achievable rate regions. Later, Han and Costa [] presented a new coding theorem for transmission of correlated sources over broadcast channels. They established a new coding scheme that specifies, instead of achievable rates, a class of source-channel matching conditions between the source and the channel. They also gave an interesting example, which reveals that separate source and channel coding is not optimal for the transmission of correlated sources over broadcast channels. The essence of Shannon s separation approach in the pointto-point case is to have an efficient architecture for transmission problems by having a discrete interface (a finite set) to represent information sources. The fundamental concept which facilitates this is the notion of typicality, which states that when grouped into large blocks, only those sequences, called typical sequences, coming from a set of size nearly equal to nh(s) are observed most of the time, and each such sequence This work was supported by NSF CAREER Grant CCF-085. has nearly the same probability, where H(S) denotes Shannon entropy []. This notion of typicality can be extended to multiple sources (say S and T ) called joint typicality. It says that even though there are roughly n(h(s)+h(t )) individually typical sequence pairs, most of the time only jointly typical sequence pairs with size nearly equal to nh(s,t ) are observed with high probability. Further, using these ideas, it can be shown that for every typical sequence of S (respectively T ), there exist roughly nh(t S) (respectively nh(s T ) ) typical T (respectively S) sequences that are jointly typical, where H(T S) is the conditional entropy [] of T given S. This leads us naturally to consider a bipartite undirected graph on the sets of individually typical sequences induced from the property of joint typicality. That is, the vertexes of this graph denote the individually typical sequences, and the jointly typical sequences are connected through an edge. We refer to this graph as the typicality-graph of two correlated sources. Based on this observation, a graph-based framework for transmission of correlated sources over multiple-access channels was considered in our earlier work [5], [6]. In essence, the transmission system would involve two modules: a source coding module and a channel coding module. In the source coding module, the sources are represented using a bipartite graph, and in the channel channel coding module, the edges coming from this graph are reliably transmitted over a multiple-access channel. In this work we consider this approach for the broadcast channels. Our goal is to devise a modular framework based on bipartite graphs for transmission of correlated sources over broadcast channels as shown in Fig.. In this transmission system, we have a source coding module and a channel coding module. In the source coding module, the sources are efficiently represented using a bipartite graph, and in the channel coding module, the edges coming from this graph are reliably transmitted over the broadcast channels. The channel coding part was studied in our earlier work in [7], where the rate of growth of the exponent (as a function of the number of channel uses) of the size of the bipartite graphs whose edges can be reliably transmitted over a broadcast channel was characterized in an information-theoretic setting. In the present work, we provide an information theoretic

characterization of the the rate of growth of the exponent (as a function of the number of source samples) of the size of the bipartite graphs which can reliably represent a pair of correlated sources to be transmitted over a broadcast channel. In the conventional separate source and channel coding scheme the correlated sources are first represented into one common message W 0 and two private messages W and W in the source coding module, and then (W 0, W ) and (W 0, W ) are reliably transmitted to the receiver and the receiver, respectively in the channel coding module. Note that here the private messages W 0, W and W are regarded as independent. Encoder Fig.. Encoder Broadcast Decoder Decoder Decoder Decoder The broadcast channel communication system The outline of the remaining part of this paper is as follows. In Section II, we provide definitions of graphs of our interest. Then, we discuss the main result of this paper and its proof in Section III and Section IV, respectively. After that, some interpretations of correlated sources are given in Section V. Finally, Section VI provides concluding remarks. II. PRELIMINARIES Before we discuss the main problem, let us first define a bipartite graph and related mathematical terms which will be used in our discussion. Definition : A bipartite graph G is defined as G = (A, A, B) where A and A are two non-empty sets of vertices, and B is a set of edges where every edge of B joins a vertex in A to a vertex in A. In general, B A A. If G is a bipartite graph, V (G) and V (G) are the first (left) and the second (right) vertex sets of G and E(G) is the edge set of G. G is said to be a complete if E(G) = V (G) V (G). The degree of a vertex v V i (G) in a graph G, denoted by deg G,i (v) is the number of edges incident to v for i =,. A subgraph of a graph G is a graph whose vertex and edge sets are subsets of those of G Since we consider particular type of bipartite graphs in our discussion, let us define those bipartite graphs. Definition : A bipartite graph G is called semiregular [8] with parameters (,,, ), denoted by G(,,, ), if it satisfies: V i (G) = i for i=,, u V (G), deg G, (u) =, v V (G), deg G, (v) =. Definition : A bipartite graph G is called nearly semiregular with parameters ( 0,,,,, µ), denoted by G( 0,,,,, µ), if it satisfies: G is composed of 0 disjoint subgraphs G m for m =,,..., 0, m, V i ( G m ) = {(m ) i +, (m ) i +,..., m i } for i=,, V i (G) = 0 m= V i( G m ) for i=,, E(G) = 0 m= E( G m ), u V (G), µ deg G, (u) µ, v V (G), µ deg G, (v) µ. Figure illustrates an example of bipartite graph G(,,,,, ) composed of two disjoint subgraphs. Definition : A triple of messages (random variables) W 0, W and W with message sets W i = {,,..., i } for i = 0,,, respectively, is said to be the message-graph, denoted by MG(G), associated with a bipartite graph G with parameters ( 0,,,,, µ) if the following condition is satisfied: w i W i, for i =,,, P r{(w 0, W, W ) = (w 0, w, w )} = E(G) if ((w 0 ) + w, (w 0 ) + w ) E(G) and P r{(w 0, W, W ) = (w 0, w, w )} = 0 else. W 0 is referred to as common message and W and W are referred to as private messages. In general, W 0 and (W, W ) are not independent (See Figure ). 5 6 7 8 Receiver Messages 5 6 7 8 Receiver Messages Fig.. Message-graph associated with G(,,,,, ) composed of two disjoint subgraphs G and G where W 0 = {, }, W = W = {,,, } III. SUMMARY OF RESULTS In broadcast channels, the source encoder can have access to both sources to be transmitted, but the source decoders can not collaborate with each other. So, this source coding problem can be called distributed reconstruction source coding problem since source decoders reconstruct transmitted sources in a distributed manner. We consider two correlated sources S and T with a joint probability distribution p(s, t) and the alphabets are given by finite sets S and T, respectively. Let (S, T ), (S, T ),... be a sequence of jointly distributed random variables i.i.d. p(s, t). For ease of exposition, we assume that the sources do not have a common part [9], [0], []. The goal is to represent these sources into message sets which can be associated with nearly semi-regular graphs with parameters ( 0,,,,, µ). Definition 5: An (n, τ)-transmission system for a bipartite graph G( 0,,,,, µ) and a pair of correlated sources (S, T ) is an ordered tuple (f, g, g ), consisting of one encoding mapping f and two decoding mappings g and g where ) f : S n T n E(G), ) g : V (G) S n, g : V (G) T n,

) such that a performance measure given by the probability of error satisfies: τ =P r{(g (f(s n,t n )), g (f(s n,t n ))) (S n, T n )}. () Now we define achievable rates as follows. Definition 6: A tuple of rates (R 0, R, R, R, R ) is said to be achievable for a source coding problem with correlated sources (S, T ) (for transmission over broadcast channels), if for any > 0, and for all sufficiently large n, there exists a bipartite graph G with parameters ( 0,,,,, µ) and an associated (n, τ)-transmission system as defined above satisfying: n log i < R i + for i = 0,,, n log i < R i + for i =,, n log µ < and the corresponding probability of error τ <. The goal is to find the achievable rate region R DR which is the set of all achievable tuple of rates (R 0, R, R, R, R ). We have found an inner bound to the achievable rate region. It is the main result of this paper, and is given by the following theorem. Theorem : R DR R DR where R DS = {(R 0, R,R, R, R ) : p(w s,t) R 0 I(S, T ; W ), () R H(S W ), () R H(T W ), () R H(S T, W ), (5) R H(T S, W ), (6) R + R = R + R H(S, T W )}, (7) where W is an auxiliary random variable such that p(w, s, t) = p(w)p(s, t w). IV. PROOF OF THEOREM In this section, we present the proof of the main result. We use the random binning technique used by Berger [], and the notion of strongly jointly typical sequences. Let us consider a fixed joint probability distribution p(w, s, t) on W S T. Also fix > 0 and real numbers R 0, R, R, R, and R. Without loss of generality, let us assume that R i < H(S, T W ) for i =,. Let R = H(S T, W ), and R = H(T S, W ). ) Random Sequences and Bin generation: First, draw nr0 sequences Wm, n m {,,..., nr0 }, of length n, independently from the strongly -typical set A (n) if w n A (n) is, P {Wm n = wn } = A (n) (W ) P {Wm n = wn } = 0 if w n / A (n) (W ). That (W ), and (W ). Then, let us define α = n(h(s) R+γ) and β = n(h(t ) R+γ) where γ will be specified shortly. Draw α sequences S n of length n independently and equally likely with replacement from the strongly -typical set A (n) (S) with cardinality n(h(s) ()) A (n) (S) n(h(s)+()) for large n where () 0 as 0 []. Then, put the all selected α sequences into a bin named B. Repeat the same procedure nr times independently, resulting in nr bins denoted by B i for i {,,..., nr }. Similarly, generate nr bins denoted by C j for j {,,..., nr }, each bin contains β sequences T n, from the strongly -typical set A (n) (T ) with cardinality n(h(t ) ()) A (n) n(h(t )+()) for large n. Here, we assume γ > (). (T ) ) Graph generation: For all (S n, T n ) A (n) (S, T ), find an index m, for m {,,..., nr0 }, such that (S n, T n, Wm) n A (n) (S, T, W ). If there is no such index m, assign m = 0. Also, find a pair of indices (i, j) where i and j are the least indices of bins such that S n B i and T n C j. If there is no bin containing S n (or T n ), assign i = 0 (or j = 0). Now we can associate a bipartite graph G with the indices m, i, and j where: G is composed of nr0 disjoint subgraphs G m for m {,,..., nr0 }, V i (G)= nr 0 m= V i( G m ) for i =,, E(G) = nr 0 m= E( G m ), m {,,..., nr0 }, V i ( G m )={(m ) nri +, (m ) nri +,..., m nri } for i =,, m {,,..., nr0 }, ((m ) nr +i, (m ) nr + j) V ( G m ) V ( G m ), ((m ) nr +i, (m ) nr + j) E( G m ) if (S n, T n ) (B i C j ) A(S, T W n m) where A(S, T W n m) is the set of (S n, T n ) such that (S n, T n, W n m ) A(n) (S, T, W ). For a particular m, a subgraph G m can be generated as shown in Figure. Fig.. Generation of a subgraph G m in a graph G from the bin indices of random sequences S n and T n ) Encoding error events: Before we proceed further, let us make sure that the generated codebooks satisfy certain properties. If the vertices of G do not satisfy certain degree requirements, we may not be able to reliably represent the sources using this graph. So, an encoding error will be declared if either one of the following events occur. Choose γ > (), and such that > γ + (). The error events are defined as: E : i V (G) such that n log deg G, (i) R >, E : j V (G) such that n log deg G,(j) R >. We now show that the probability of these events can be made arbitrarily small for sufficiently large n. Let us define two events E 0, and E 0, as follows: E 0, : i V (G) such that deg G, (i) < n(h(s T,W ) ), E 0, : j V (G) such that deg G, (j) < n(h(t S,W ) ).

Lemma : Then for any > 0, and sufficiently large n: P {E 0,i } <, for i =, (8) 6 Proof: The proof of Lemma is omitted. Let us define two events E0, and E0, as follows. E0, : i V (G) such that deg G, (i) > n(h(s T,W )+ ), E0, : j V (G) such that deg G, (j) > n(h(t S,W )+ ). Lemma : Then for any > 0, and sufficiently large n: P {E0,i} <, for i =, (9) 6 Proof: The proof of Lemma is omitted. So, by Lemma and Lemma, P (E ) + P (E ) <. In other words, with high probability we can obtain a bipartite graph G where each vertex in V (G) has degree nearly equal to nh(s T,W ) and each vertex in V (G) has degree nearly equal to nh(t S,W ). ) Choosing message-graphs: If any of E or E occur, then choose any graph with parameters ( nr0, nr, nr, nr, nr, n ), and assume that the triple of messages W 0, W and W is the message-graph associated with this graph. In this case, no guarantees will be given regarding the probability of error. If none of these events occur, then choose the message-graph M G(G) as the triple of messages. 5) Encoding: For a given sources (S n, T n ), sender sends (m, i) and (m, j) to receiver and receiver, respectively. 6) Decoding: Given the received index pair (m, i), receiver declares Ŝn = S n if there exists a unique sequence S n that is jointly typical with the particular sequence Wm n and belongs to B i, i.e., S n B i A(S Wm) n where A(S Wm) n the set of S n such that (S n, Wm n ) A(n) (S, W ). Otherwise, declare an error. Similarly, given (m, j), receiver declares ˆT n = T n if there is a unique sequence T n such that T n C j A(T Wm n ) where A(T W m n) the set of T n such that (T n, Wm) n A (n) (T, W ). Otherwise, an error is declared. 7) Probability of error analysis: Let E denote the event {(g (m, i), g (m, j)) (S n, T n )}, that the reconstructed sequence pair is not equal to the original source sequence pair. The probability of error P (E) can be given by P (E) = P (E E )P (E E E )+P (E E c Ec ) (0) P (E E ) + P (E E c Ec ). () The second probability in the above equation can be bounded as given in the following lemma. Lemma : For any > 0, and sufficiently large n, provided P (E E c E c ) < () R 0 > I(S, T ; W ) + (), () R > H(S W ) + γ + (), () R > H(T W ) + γ + () (5) where () 0 as 0 and n. Proof: Now let us calculate the probability P (E E c E). c If previous error events E or E do not occur, we define other error events as follows. E : (S n, T n ) / A (n), E : m {,,..., nr0 } such that (S n, T n, Wm n ) A (n) (S, T, W ), i.e., m = 0, E 5 : i {,,..., nr } such that S n B i, i.e., i = 0, E 6 : j {,,..., nr } such that T n C j, i.e., j = 0, E 7 : S n S n such that S n B i A(S Wm), n E 8 : T n T n such that T n C j A(T Wm n). Then, P (E E c E) c = P ( 8 i=e i E c E c ) 8 i= (6) P (E i E c Ec ) (7) By the property of jointly typical sequences [], P (E E c E) c < 8 for sufficiently large n. P (E E c E c ) (8) = P [ m {,,..., nr0 }, (S n, T n, Wm n )/ A(n) (S, T, W )] (9) [ = P {(S n, T n, W n ) A(n) (S, T, W )}] nr0 (0) [ n(i(s,t ;W )+())] nr 0 () exp( n(r0 I(S,T ;W ) ) ) () < 8 () for sufficiently large n, if R 0 > I(S, T ; W ) + () where () 0 as 0 and n. Following [], for sufficiently large n P (E 5 E c Ec ) < 8 P (E 6 E c E) c < 8 if γ >. For sufficiently large n, it can be shown that () (5) P (E 7 E c Ec ) = P { S n S n : S n B i A(S W n m )} (6) (a) n(h(s) R+γ) n(i(s;w ) ) (7) = n(h(s W ) R+γ+) (8) < 8 (9) provided R > H(S W )+γ+, where (a) is from the union bound. Similarly, sufficiently large n, P (E 8 E c E c ) < 8 if R > H(T W ) + γ +. Therefore, P (E) P (E )+P (E )+P (E E0 c Ec Ec ) <. (0) Therefore P (E) < for sufficiently large n and under the conditions given by the theorem. In every realization

of random codebooks we have obtained a message-graph M G(G) and an associated graph G with the same parameters, and averaged over this ensemble, we have made sure that the probability of error is within the tolerance level of. Hence, the proof of the direct coding theorem is completed. V. DIFFERENT MESSAGGRAPHS FOR A PAIR OF CORRELATED SOURCES IN THE BROADCAST CHANNEL We have shown that in the broadcast channel, for sufficiently large block length n, a pair correlated sources (S, T ) can be represented using nearly semi-regular graph G with parameters ( nr0, nr, nr, nr, nr, n ) as shown in Figure. Fig.. Graph G with parameters ( nr 0, nr, nr, nr, nr, n ) composed of nr 0 subgraphs G m( nr, nr, nr, nr, n ) Note that many different message-graphs can represent the same correlated sources without increasing the redundancy. Let us consider a case where the sum rate is minimum, i.e., R 0 + R + R = H(S, T ). In this case, the total number of edges in the graph, E(G) = n(i(s,t ;W )+H(S,T W )) = nh(s,t ). Since we can vary the mutual information I(S, T ; W ) such that 0 I(S, T ; W ) H(S, T ), we can have many message-graphs without increasing redundancy. Let us consider some special cases as follows. Case A: If I(S, T ; W ) = 0, then R 0 = 0, R = H(S), R = H(T ), R = H(S T ), R = H(T S). Roughly speaking, this corresponds to the typicalitygraph of (S, T ). Case B: If I(S, T ; W ) = H(S, T ), then R 0 = H(S, T ), R = 0, R = 0, R = 0, R = 0. Case C: If 0 < I(S, T ; W ) < H(S, T ), then R 0 = I(S, T ; W ), R = H(S W ), R = H(T W ), R = H(S T, W ), R = H(T S, W ). In particular, consider the case when R 0 = C(S; T ), where C(S; T ) denotes the common information defined by Wyner [] as follows. C(S; T ) = inf I(S, T ; W ) where the infimum is taken over all triples of random variables S, T, W, where auxiliary random variable W take values on a finite set, and such that () the marginal distribution for S, T is p(s, t), () S and T are conditionally independent given W, i.e., S W T and p(s, t, w) = p(w)p(s w)p(t w). In this case, R 0 = C(S; T ), R = H(S W ), R = H(T W ), R = H(S W ), R = H(T W ) since S and T are conditionally independent given W. So, each subgraph G m become complete. This also means that the private messages for receiver and become independent. Note that a subgraph G m can not be complete if R 0 < C(S; T ) because, by the definition, C(S; T ) is the infimum of I(S, T ; W ) such that S W T. Hence Case A can be thought of as situated at one end of the spectrum, and Case B as situated on the other end of the spectrum. For every value of I(S, T ; W ), we get an equally efficient representation of the sources into a nearly semiregular graph. Here, equally efficient representation means that the cardinalities of edge sets of these graphs for the different values of I(S, T ; W ) are nearly the same. VI. CONCLUSION We have considered the task of mapping the outputs of the two correlated sources (for transmission over broadcast channels) into a triple of messages, i.e., a common message and a pair of correlated private messages, which are associated with an undirected nearly semi-regular bipartite graph. We have shown that a pair of correlated sources can be represented into many different triple of messages without increasing the redundancy. This study corresponds to the source coding part of our approach for the transmission of correlated sources over broadcast channels. The goal of this work is to show that a message-graph, which is a triple of messages associated with a nearly semi-regular bipartite graph, can be used as discrete interface in this Shannon-style modular approach to multiterminal communication problems. REFERENCES [] T. M. Cover, Broadcast channels, IEEE Trans. Inform. Theory, vol. IT-8, no., pp., Jan. 97. [] K. Marton, A coding theorem for the discrete memoryless broadcast channel, IEEE Trans. Inform. Theory, vol. IT-5, no., pp. 06, May 979. [] T. S. Han and H. M. Costa, Broadcast channels with arbitrarily correlated sources, IEEE Trans. Inform. Theory, vol. IT-, no. 5, pp. 6 650, Sep. 987. [] T. M. Cover and J. A. Thomas, Elements of Information Theory. New York:Wiley, 99. [5] S. S. Pradhan, S. Choi, and K. Ramchandran, Achievable rates for multiple-access channels with correlated messages, in Proc. IEEE Int. Symp. on Inform. Theory (ISIT), Jun. 00. [6], A graph-based framework for transmission of correlated sources over multiple access channel, submitted to IEEE Trans. Inform. Theory, Jan. 006. [7] S. Choi and S. S. Pradhan, Transmission of correlated messages over the broadcast channel, in Proc. Conf. on Inform. Sciences and Systems (CISS), Mar. 005. [8] J. H. van Lint and R. M. Wilson, A course in combinatorics. Cambrigde University Press, 00. [9] P. Gács and J. Körner, Common information is much less than mutual information, Problems of Control and Information Theory, vol., pp. 9 6, 97. [0] H. S. Witsenhausen, On sequences of pairs of dependent random variables, SIAM J. Appl. Math., vol. 8, pp. 00, Jan. 975. [] A. D. Wyner, The common information of two dependent random variables, IEEE Trans. Inform. Theory, vol. IT-, no., pp. 6 79, Mar. 975. [] T. Berger, Multiterminal Coding. In: The Information Theory Approach to Communications (ed. G. Longo), CISM Courses and Lecture Notes No. 9. Springer, Wien-New York, 977.