Key Capacity with Limited One-Way Communication for Product Sources

Similar documents
A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources

AN INTRODUCTION TO SECRECY CAPACITY. 1. Overview

EE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15

Performance-based Security for Encoding of Information Signals. FA ( ) Paul Cuff (Princeton University)

A Formula for the Capacity of the General Gel fand-pinsker Channel

A new converse in rate-distortion theory

Soft Covering with High Probability

Lattices for Distributed Source Coding: Jointly Gaussian Sources and Reconstruction of a Linear Function

Hypercontractivity, Maximal Correlation and Non-interactive Simulation

Achieving Shannon Capacity Region as Secrecy Rate Region in a Multiple Access Wiretap Channel

Capacity Region of Reversely Degraded Gaussian MIMO Broadcast Channel

Dispersion of the Gilbert-Elliott Channel

On the Simulatability Condition in Key Generation Over a Non-authenticated Public Channel

On Function Computation with Privacy and Secrecy Constraints

Arimoto Channel Coding Converse and Rényi Divergence

SHARED INFORMATION. Prakash Narayan with. Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe

Convexity/Concavity of Renyi Entropy and α-mutual Information

An Alternative Proof for the Capacity Region of the Degraded Gaussian MIMO Broadcast Channel

The Gallager Converse

Channels with cost constraints: strong converse and dispersion

Solutions to Homework Set #2 Broadcast channel, degraded message set, Csiszar Sum Equality

Remote Source Coding with Two-Sided Information

SHARED INFORMATION. Prakash Narayan with. Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe

Lecture 4 Noisy Channel Coding

Secret Key Agreement Using Asymmetry in Channel State Knowledge

arxiv: v4 [cs.it] 17 Oct 2015

Stochastic Realization of Binary Exchangeable Processes

Common Randomness and Secret Key Generation with a Helper

On Gaussian MIMO Broadcast Channels with Common and Private Messages

Group Secret Key Agreement over State-Dependent Wireless Broadcast Channels

Secret Key Agreement: General Capacity and Second-Order Asymptotics. Masahito Hayashi Himanshu Tyagi Shun Watanabe

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Secret Key Agreement Using Conferencing in State- Dependent Multiple Access Channels with An Eavesdropper

Degrees of Freedom Region of the Gaussian MIMO Broadcast Channel with Common and Private Messages

Two Applications of the Gaussian Poincaré Inequality in the Shannon Theory

Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels

Perfect Omniscience, Perfect Secrecy and Steiner Tree Packing

PERFECTLY secure key agreement has been studied recently

ProblemsWeCanSolveWithaHelper

Katalin Marton. Abbas El Gamal. Stanford University. Withits A. El Gamal (Stanford University) Katalin Marton Withits / 9

Amobile satellite communication system, like Motorola s

On Scalable Coding in the Presence of Decoder Side Information

Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets

The Poisson Channel with Side Information

Fading Wiretap Channel with No CSI Anywhere

On Common Information and the Encoding of Sources that are Not Successively Refinable

SUCCESSIVE refinement of information, or scalable

Secret Key Generation and Secure Computing

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information

UCSD ECE 255C Handout #12 Prof. Young-Han Kim Tuesday, February 28, Solutions to Take-Home Midterm (Prepared by Pinar Sen)

Secret Key and Private Key Constructions for Simple Multiterminal Source Models

Reliable Computation over Multiple-Access Channels

ON THE BOUNDEDNESS BEHAVIOR OF THE SPECTRAL FACTORIZATION IN THE WIENER ALGEBRA FOR FIR DATA

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016

Non-interactive Simulation of Joint Distributions: The Hirschfeld-Gebelein-Rényi Maximal Correlation and the Hypercontractivity Ribbon

Shannon s A Mathematical Theory of Communication

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

Secret Key Agreement: General Capacity and Second-Order Asymptotics

Channel Dispersion and Moderate Deviations Limits for Memoryless Channels

Coordination Capacity

Entropy and Ergodic Theory Notes 21: The entropy rate of a stationary process

Simple Channel Coding Bounds

Upper Bounds on the Capacity of Binary Intermittent Communication

A New Metaconverse and Outer Region for Finite-Blocklength MACs

Ahmed S. Mansour, Rafael F. Schaefer and Holger Boche. June 9, 2015

Multi-Kernel Polar Codes: Proof of Polarization and Error Exponents

EECS 750. Hypothesis Testing with Communication Constraints

ELEC546 Review of Information Theory

Information Masking and Amplification: The Source Coding Setting

Latent Tree Approximation in Linear Model

Information Theoretic Limits of Randomness Generation

An Alternative Proof of Channel Polarization for Channels with Arbitrary Input Alphabets

Chain Independence and Common Information

Refined Bounds on the Empirical Distribution of Good Channel Codes via Concentration Inequalities

Chapter I: Fundamental Information Theory

5958 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 12, DECEMBER 2010

Research Summary. Chandra Nair. Overview

(Classical) Information Theory III: Noisy channel coding

Common Randomness Principles of Secrecy

The Least Degraded and the Least Upgraded Channel with respect to a Channel Family

0.1 Uniform integrability

Intermittent Communication

A New Achievable Region for Gaussian Multiple Descriptions Based on Subset Typicality

On Capacity Under Received-Signal Constraints

Side-information Scalable Source Coding

The Shannon s basic inequalities refer to the following fundamental properties of entropy function:

MULTITERMINAL SECRECY AND TREE PACKING. With Imre Csiszár, Sirin Nitinawarat, Chunxuan Ye, Alexander Barg and Alex Reznik

Topological properties

Optimal Natural Encoding Scheme for Discrete Multiplicative Degraded Broadcast Channels

Variable Length Codes for Degraded Broadcast Channels

On the Rate-Limited Gelfand-Pinsker Problem

Functional Properties of MMSE

Duality Between Channel Capacity and Rate Distortion With Two-Sided State Information

Block 2: Introduction to Information Theory

The Capacity Region of a Class of Discrete Degraded Interference Channels

Gaussian Estimation under Attack Uncertainty

Graph Coloring and Conditional Graph Entropy

INFORMATION-THEORETIC BOUNDS OF EVOLUTIONARY PROCESSES MODELED AS A PROTEIN COMMUNICATION SYSTEM. Liuling Gong, Nidhal Bouaynaya and Dan Schonfeld

SOURCE CODING WITH SIDE INFORMATION AT THE DECODER (WYNER-ZIV CODING) FEB 13, 2003

Approximately achieving the feedback interference channel capacity with point-to-point codes

Transcription:

Key Capacity with Limited One-Way Communication for Product Sources Jingbo Liu Paul Cuff Sergio Verdú Dept. of Electrical Eng., Princeton University, NJ 08544 {jingbo,cuff,verdu}@princeton.edu ariv:1405.7441v1 [cs.it] 29 May 2014 Abstract We show that for product sources, rate splitting is optimal for secret key agreement using limited one-way communication at two terminals. This yields an alternative proof of the tensorization property of a strong data processing inequality originally studied by Erkip and Cover and amended recently by Anantharam et al. We derive a water-filling solution of the communication-rate key-rate tradeoff for two arbitrarily correlated vector Gaussian sources, for the case with an eavesdropper, and for stationary Gaussian processes. I. INTRODUCTION The fundamental limit on the amount of secret key (or common randomness that can be generated by two terminals which observe correlated discrete memoryless sources was studied in [1],[2], where single-letter solutions were derived for the class of protocols with one-way communication. However, in many practical applications, the key capacity is not known, since the optimizations over auxiliary random variables in those single-letter formulas are usually hard to solve. In [3] the fundamental limit was extended to sources with continuous alphabets, and it was shown that for vector Gaussian sources it suffices to consider auxiliary random variables which are jointly Gaussian with the sources. Thus the capacity region for vector Gaussian sources was posed as a matrix optimization problem. Still, an explicit formula for the key capacity was not derived except for scalar Gaussian sources. In this paper we provide the explicit formula for key capacity of vector Gaussian sources by considering what turns out to be more general: the key capacity of arbitrary product sources. Specifically, suppose terminals A, B and an eavesdropper observe discrete memoryless vector sources ( i n, Y (Y i n and Z (Z i n respectively, where n n P Y P iy i, P Z P iz i. (1 We call a source of this kind a product source because of the structure of its joint probability distribution. The maximal rate of secret key achievable as a function of public communication rate r from A to B is denoted as R(r. We show that R(r max r i r R i (r i, (2 where R i (r i is the maximal achievable key rate given a communication rate of r i for the i th source triple: ( i,y i,z i. This is analogous to a result in rate distortion theory about rate for a product source with a sum distortion measure [4], where Shannon showed that the rate distortion function is derived by summing the rates and distortions of points in the curves with the same slope. In the case of jointly vector Gaussian sources without an eavesdropper (or with an eavesdropper but under suitable conditions, one can always apply separate invertible linear transforms on the vectors observed at A and B so that the source distribution is of the form in (1, thus deriving an explicit formula of R(r utilizing corresponding results of scalar Gaussian sources. The solution displays a water filling behavior similar to the rate distortion function of vector Gaussian sources [5]. We shall also discuss the extension to stationary Gaussian processes. There is a curious connection between our result about R(r for product sources and the tensorization property of a strong data processing inequality originally studied by Erkip and Cover and amended recently by Anantharam et al. [6]. Suppose P Y is given. In [7] it was mistakenly claimed that s I(U;Y (;Y : sup U Y,I(U; 0 I(U; ρ2 m(;y, where ρ 2 m (;Y denotes the maximal correlation coefficient [8]. In fact, [6] shows that this is not always true and gives a general while less explicit simplification for s (;Y: s (;Y sup Q P D(Q Y P Y D(Q P. (3 Nevertheless, ρ 2 m(;y and s (;Y do agree for some simple distributions of P Y such as Gaussian and binary with equiprobable marginals, and they both fulfill an important property called tensorization. Moreover, they are closely linked to the problem of key generation [8][9]. 1 To add one more connection between s (;Y and key generation, we demonstrate that (2 implies the tensorization property of s (;Y. II. PRELIMINARIES Consider the model explained in Section I with the sources distributed as in (1. We use N to denote block length since n has been reserved for the length of the vector in (1. The standard notation N 1 N will be used for the block ( i N. Upon receiving N, terminal A computes an integer K 1 K 1 ( N K 1 and sends a message W W( N W through a noiseless public channel to terminal B. Then B computes the keyk 2 K 2 (W( N,Y N based on available 1 For the reason we just discussed, the ρ 2 m (;Y in the expressions of efficiency functions in [9] should be replaced by s (;Y.

information. A rate pair (r, R is said to be achievable if it can be approached by a sequence of schemes satisfying the following conditions on the probability of agreement and security: The function lim 1 K 2 0, N (4 lim 1 H(K 1 W,Z N ] 0, N (5 1 lim sup log W r, N N (6 1 lim inf N N log K 1 R. (7 R(r : sup{r : (r,r is achievable} (8 characterizes the maximal possible key rate with a certain public communication rate. III. MAIN RESULTS A. Initial Efficiency in Key Generation Fix P YZ. Define s Z(;Y sup U,V I(V; U I(V;Z U+I(U; I(U;Y, (9 where the supremum is over all (U, V such that (U, V (Y,Z and that the denominator in (9 doesn t vanish. Note that the denominator is always nonnegative; if it vanishes for all U,V, then so does the numerator and we set s Z (;Y 0. s Z (;Y is closely related to the initial efficiency of key R(r generation, defined as lim r 0 r. In the special case of no eavesdropper, this is related to the result in [9], which uses the incorrect constant ρ 2 m (;Y as we mentioned earlier. The following result gives some basic properties of s Z (;Y. Theorem 1. 1 0 s Z (;Y 1. 2 s Z (;Y tensorizes, in the sense that for product sources, s Z n(n ;Y n max 1 i n s Z i ( i ;Y i. (10 3s Z (;Y is linked to the initial efficiency of key generation by 4 s Z(;Y R(r R(r lim sup r 0 r r>0 r s Z (;Y 1 s (11 Z (;Y. I( V;Ȳ I( V; sup Z Q V I( V; I( V; Z+D(Q P D(Q Y P Y, (12 where V,,Ȳ, Z has joint distribution P V Ȳ Z(v,x,y,z Q V (v,xp Y Z (y,z x. The supremum is over all Q V such that the above denominator does not vanish. Computation can be further simplified for degraded sources Y Z: s I(U;Y Z Z (;Y sup U I(U; Z where Q YZ Q P YZ. sup Q D(Q Y P Y D(Q Z P Z D(Q P D(Q Z P Z (13 Proof: 1 From the data processing inequality the denominator in (9 is nonnegative, and s Z (;Y 1. If there exists U such that I(U; I(U;Y > 0, we can choose V independent of U,, Y so that the numerator vanishes whereas the denominator is positive, which shows that s Z (;Y 0. Otherwise if I(U; I(U;Y 0 for all U, the numerator will always be nonnegative: I(U,V;Y I(U,V;Z I(U;Y+I(U;Z I(U,V; I(U,V;Z I(U;+I(U;Z. (14 Hence s Z (;Y 0 always holds. 2 We only need to show that s Z (n ;Y n max 1 i n s Z i ( i ;Y i since the other direction is obvious. For any U,V such that (U,V 1 n (Y1 n,z1 n and I(U,V;1 n I(U,V;Y 1 n > 0,I(V;Y n U I(V;Z n U > 0, let U1 n,v 1 n be as in Lemma 2 in the appendix. Then I(V;Y n U I(V;Z n U I(U,V; n I(U,V;Y n [I(V i;y i U i I(V i ;Z i U i ] [I(U i,v i ; i I(U i,v i ;Y i ] I(V i ;Y i U i I(V i ;Z i U i max i I I(U i,v i ; i I(U i,v i ;Y i max sup I(V i ;Y i U i I(V i ;Z i U i 1 i nu i,v i I(U i,v i ; i I(U i,v i ;Y i (15 (16 s Z (;Y 1 s Z (;Y, which is equivalent to the where I is the set of indices such that I(U i,v i ; i I(U i,v i ;Y i 0, and the suprema are over allu i,v i such that (U i,v i i (Y i,z i and I(U i,v i ; i I(U i,v i ;Y i 0. Supremizing with respect to U, V on (15 shows the tensorization property of tensorization property of s Z (;Y. 3 The achievable region of (r,r is the union of [I(U,V; I(U,V;Y, [0,] (20 over all U,V such that (U,V (Y,Z [1]. Thus R(r sup r>0 r s Z (;Y 1 s Z (;Y follows immediately from the definition of s Z (;Y. The claim of lim r 0 R(r R(r r sup r>0 r follows from the convexity of the achievable rate region. 4 One direction of the inequality is obvious: if U,V are such that I(V;Y U I(V;Z U 0, we have (see (17- (19.

Conversely, for any Q V, consider Q (1 VYZ Q VP Y Z, (21 Q (0 VYZ P V PYZ αq (1 YZ, (22 1 α PY α ZUV(x,y,z,u,v (1 αq (0 V YZ (v,x,y,z1 u0 +αq (1 V YZ (v,x,y,z1 u1, (23 where P V is an arbitrary probability distribution on V. We can assume that P YZ (x,y,z > 0 for all x,y,z so that (22 is a well-defined distribution for α > 0 small enough. Then, we can verify that P α YZ P Y Z, (U,V (Y,Z, and lim α 0 I(V; U I(V;Z U+I(U; I(U;Y I( V;Ȳ I( V; Z I( V; I( V; Z+D(Q P D(Q Y P Y, (24 where P UV YZ : PUV α YZ. Thus (12 holds. The proof of (13 is omitted here due to space constraint. Thanks to Theorem 1, we can always eliminate one auxiliary r.v. if we only want to compute s Z (;Y instead of the rate region, which considerably reduces the dimension in the optimization problem. The interpretation of the tensorization of s Z (;Y is that, with small allowable public communication, it is always efficient to only use the best component of the product sources. Alternatively, the fact that rate splitting is optimal for product sources (as we shall prove in the sequel implies the tensorization property of s Z (;Y. B. Secret Key Generation from Product Sources We show that by appropriately splitting the communication rate to each factor in the product source, producing keys separately and combining them (called rate splitting, one can always achieve the optimal key rate. Theorem 2. In the problem of key generation from product sources satisfying (1, the maximum key rate satisfies R(r max r i r R i (r i, (25 which can be achieved by rate splitting. HereR i (r i is the keyrate communication-rate function corresponding to the i th source triple ( i,y i,z i. Further, if R i ( is differentiable and R(r R i (ri, (26 then for each i, either R i (r i µ or r i 0, where µ is some constant. Proof: Each rate of R i (ri can be approached by a scheme that operates on each of the i th source triple separately. From the second equation in (1, the combination of these schemes form a legitimate scheme for the product source. Thus, the direction of R(r R i(r i is trivial. By (20 the achievable region of (r,r is the union of [I(U,V; n 1 I(U,V;Y n 1, [0,I(V;Y n 1 U I(V;Z n 1 U] (27 over all (U,V such that (U,V 1 n (Y1 n,z1. n The achievable region with rate splitting is the union of [ I(U i,v i ; i I(U i,v i ;Y i, [ 0, I(V i ;Y i U i ] I(V i ;Z i U i (28 over all (U i,v i such that (U i,v i i (Y i,z i. The second union contains the first union, according to result of Lemma 2 in the appendix. Hence we also have R(r R i(ri for some ri, i 1...n. The last claim in the theorem for differentiable R i ( is from the KKT condition. As an application, we derive a water-filling solution for the communication-rate key-rate tradeoff from two arbitrarily correlated vector Gaussian sources. For a nonnegative definite matrix Σ, let Σ 1/2 be a positive definite matrix such that Σ 1/2 ΣΣ 1/2 I r, where I r denotes the identity matrix of dimension r rank(σ. Also write Σ 1 (Σ 1/2 2, which agrees with the definition of inverse matrix when Σ is invertible. Note that Σ 1/2 (and therefore Σ 1 may not be unique, although their choices do not affect the value of our key capacity expression. The following fact about Gaussian distributions is useful. The proof is based on singular value decomposition and is omitted here due to limitation of space. Lemma 1. Suppose, Y, Z are jointly Gaussian vectors. There exist invertible linear transforms, Y Ȳ, Z Z such that all the five covariance I(V; U I(V;Z U+I(U; I(U;Y [I(V;Y U u I(V;Z U u]dpu (u [I(V; U u I(V;Z U u+d(p Uu P D(P Y Uu P Y ]dp U (u (17 sup u I(V;Y U u I(V;Z U u I(V; U u I(V;Z U u+d(p Uu P D(P Y Uu P Y I( V;Ȳ I( V; sup Z Q V I( V; I( V; Z+D(Q P D(Q Y P Y. (19 (18

matrices Σ, Σ Ȳ, Σ Z, Σ Ȳ, Σ Z are diagonalized if and only if Σ 1/2 Σ ZΣ 1 Z Σ ZΣ 1/2 commutes with Σ 1/2 Σ YΣ 1 Y Σ YΣ 1/2. Remark 1. Luckily, the commutativity assumption is satisfied by stationary Gaussian processes (asymptotically over large blocklengths. This is due to the commutativity of convolution. Corollary 1. If, Y are jointly Gaussian vectors, then there exist invertible linear transforms, Y Ȳ such that Σ, Σ Ȳ, Σ Ȳ are diagonalized. Thanks to Theorem 2 and Corollary 1, the task of finding the key capacity of arbitrarily correlated Gaussian vector sources in the absence of eavesdroppers is reduced to the case of product Gaussian sources ( n,y n satisfying (1. In the presence of an eavesdropper, it is not always possible to reduce the problem to the case of product sources, since the conditions in Theorem 2 are not always fulfilled; but we discuss its practical relevance later. We now present the key capacity of product Gaussian sources. The solution displays a water-filling behaviour which is reminiscent of the ratedistortion function for Gaussian vectors [5]. Theorem 3. If ( n,y n,z n are product Gaussian sources, then the achievable communication and key rates are parameterized by µ > 0 as r 1 2 R 1 2 where β i : ρ2 i Y i ρ 2 i Z i 1 ρ 2 i Y i. i:β i>µ i:β i>µ log β i(µ+1 (β i +1µ, (29 log β i +1 µ+1, (30 Remark 2. The initial efficiency is max 1 i n β + i, where β + i : max{β i,0}. Proof: Reference [3] derived an explicit formula for the achievable key rate in the case of degraded scalar sources: R(r 1 2 log Σ y xz2 2r +Σ y z (1 2 2r Σ y xz (31 1 2 log ( 1 ρ 2 xz 1 ρ 2 xy + ρ2 xy ρ2 xz 1 ρ 2 2 2r. (32 xy Note that the achievable rate region depends only on the marginal distributions P Y and P Z (see (20, hence (32 holds as long as ρ xy ρ xz, in which case,y,z is stochastically degraded [5] in that order. When ρ xy < ρ xz, from (27 we see that R(r 0 since, Z, Y is stochastically degraded in that order. Hence in all cases, we can further simplify R(r 1 2 log( 1+β + β + 2 2r (33 where β : ρ2 xy ρ2 xz 1 ρ and recall the notation β + : 2 xy max{β,0}. Now consider the product sources, and suppose ri are the numbers that achieve the maximum in Theorem 2. According to Theorem 2, either R i (r i β + i 2 2r i 1+β + i β+ i 2 2r i µ or r i 0 for each i, where µ is some constant. For fixed µ, this means ri {0, max 1 } 2 log (1+µβ+ i µ(1+β i +. (34 Equivalently, we can write r i 1 2 log β+ i (m i +1 (β + i +1m i, (35 where m i : min{µ,β i + }. The claim then follows by substituting the value of ri into (33 and applying (25. Theorem 3 can be used to derive the key-rate communication-rate tradeoff for Wiener class 2 stationary Gaussian processes, in which case the linear transforms in Lemma 1 can be easily found. We shall only discuss the basic idea here since a rigorous derivation is complicated and involves the asymptotic behaviour of Toeplitz matrices. Consider first passing Y through a filter whose impulse response is R Y, 3 the correlation function between and Y, resulting in a new process Ŷ. Similarly, construct Ẑ by convolving with R Z. Note that R Ŷ R Y R Y, (36 R Ẑ R Z R Z, (37 become symmetric functions. Setting Q n, Ȳ QŶ n, Z QẐn where Q the sine/cosine orthogonal matrix, the covariance matrices Σ, ΣȲ, Σ Z, Σ Ȳ, Σ Z will be approximately diagonal (asymptotically for large n. In summary, in the spectral representation the original Gaussian sources are converted sources satisfying the product assumption (1, and the correlation coefficients corresponding to frequency ω are ρ Y (ω S Y(ω S (ωs Y (ω, (38 ρ Z (ω S Z(ω S (ωs Z (ω, (39 where S,S Y,S Z,S Y,S Z denote the spectral densities and joint spectral densities. Using (38, (39 and Theorem 3, the key capacity of Gaussian processes is obtained as follows: Theorem 4. For Wiener class stationary Gaussian processes, we have r 1 log β(ω(µ+1 dω, (40 4π β(ω>µ (β(ω+1µ R 1 4π β(ω>µ log β(ω+1 dω. (41 µ+1 where β(ω SY (ω 2 S Z(ω S Z(ω 2 S Y (ω S (ωs Y (ωs Z(ω S Y (ω 2 S Z(ω. Details of the proof are omitted here. Remark 3. Initial efficiency is the essential supremum of β +. 2 We say a Gaussian process is in the Wiener class if its correlation function is absolutely summable [10]. This is merely a technical condition which facilitates the proof. 3 When R Y is bandlimited, convolution with R Y becomes a degenerate linear transform. In this case we can use a signal ˆRY as an alternative, where ˆR Y has full spectrum and agrees with R Y in the pass-band of R Y. The final formula of key capacity however will remain unchanged.

IV. DISCUSSION The evidence points to the principle that rate splitting is optimal for product resources asymptotically in most coding problems admitting single-letter information theoretic solutions. Indeed, the algebraic manipulations in the converse proofs usually rely only on the independence of { t }, rather than that they are identically distributed. Hence the main element in proving such a result about rate splitting (e.g. Lemma 2 in the appendix is usually related to the converse proof of the corresponding coding theorem. However, this is not the case for coding problems of combinatorial nature. For example, the zero-error capacity of independent channels operating in parallel is not the sum of the zero-error capacities of the individual channels [11]. ACKNOWLEDGMENT This work was supported by NSF under Grants CCF- 1116013, CCF-1319299, CCF-1319304, and the Air Force Office of Scientific Research under Grant FA9550-12-1-0196. APPENDI [I(U,V,Y i 1, Z i+1 n ; i I(U,V,Y i 1, Z i+1 n ;Y i] + [I(i+1 n ; i U,V,Y i 1, Z i+1 n I(i+1 n ;Y i U,V,Y i 1, Z i+1 n ] (49 [I(U,V,Y i 1, Z i+1 n ; i I(U,V,Y i 1, Z i+1 n ;Y i] (50 [I(Ūi,V i ; i I(Ūi,V i ;Y i ]. (51 Here, (46 is again by [12, Lemma 4.1], and (47 is from ( i,y i (i+1 n,y i 1. Equality (48 is from Z i+1 n (U,V,n i+1,y i 1 ( i,y i. Inequality (50 is because of I(i+1 n ; i U,V,Y i 1, Z i+1 n I(i+1 n ; i,y i U,V,Y i 1, Z i+1 n. Finally choose a random vector U n satisfying P Ui V i iy iz i (u i v i,x i,y i,z i P Ūi V i i (u i v i x i. (52 Lemma 2. Suppose that {( i,y i,z i } n possess the product structure of (1, and (U,V are r.v. s such that (U,V Since P UiV i iz i P ŪiV i i Zi, we have 1 n (Y 1 n,zn 1. Then there exist Un 1 and V1 n such that I(V i ;Z i U i I(V i ; Z i Ūi. (53 (U i,v i i (Y i,z i for i 1,...,n and By the same token, we also have n I(U,V;1 n I(U,V;Y 1 n [I(U i,v i ; i I(U i,v i ;Y i ], I(V i ;Y i U i I(V i ;Y i Ūi, I(U i,v i ;Y i I(Ūi,V i ;Y i, (54 (42 I(U i,v i ; i I(Ūi,V i ; i, I(V;Z1 U n I(V; Z 1 U. n (55 I(V;Y1 n U I(V;Z1 U n [I(V i ;Y i U i I(V i ;Z i U i ]. The last five identities and (45, (51 complete the proof. (43 Proof: Choose a random vector Z REFERENCES n such that [1] I. Csiszár and P. Narayan, Secrecy capacities for multiple terminals, P UV n Y n Z n(u,v,xn,y n,z n IEEE Transactions on Information Theory, vol. 50, no. 12, pp. 3047 3061, 2004. P UV n(u,v,x n P Y n n(yn x n P Zn n(zn x n. (44 [2] R. Ahlswede and I. Csiszár, Common randomness in information theory and cryptography. ii. cr capacity, IEEE Transactions on Information Define Ūi (Y i 1, Z i+1 n,u and V i V. Then (Ūi,V i Theory, vol. 44, no. 1, pp. 225 240, 1998. i (Y i, Z [3] S. Watanabe and Y. Oohama, Secret key agreement from vector i forms a Markov chain for each i. Moreover gaussian sources by rate limited public communication, Information Theory Proceedings (ISIT, 2010 IEEE International Symposium on, I(V;Y1 n U I(V; Z 1 U n [I(V i ;Y i Ūi I(V i ; Z pp. 2597 2601, 2010. i Ūi] [4] C. E. Shannon, Coding theorems for a discrete source with a fidelity criterion, IRE Nat. Conv. Rec, vol. 4, no. 142-163, 1959. (45 [5] T. M. Cover and J. A. Thomas, Elements of Information Theory. John Wiley & Sons, 2012. holds, c.f. [12, Lemma 4.1]. Next, observe that [6] V. Anantharam, A. Gohari, S. Kamath, and C. Nair, On maximal correlation, hypercontractivity, and the data processing inequality studied I(U,V;1 I(U,V;Y n 1 n by erkip and cover, ariv preprint ariv:1304.6133, 2013. [7] E. Erkip and T. M. Cover, The efficiency of investment information, [I(U,V; i i+1 n,y i 1 I(U,V;Y i i+1 n Y i 1 IEEE Transactions on Information Theory, vol. 44, no. 3, pp. 1026 ] 1040, 1998. [8] H. S. Witsenhausen, On sequences of pairs of dependent random (46 variables, SIAM Journal on Applied Mathematics, vol. 28, no. 1, pp. 100 113, 1975. [I(U,V,i+1 n,y i 1 ; i I(U,V,i+1 n Y i 1 ;Y i ] [9] L. Zhao, Common randomness, efficiency, and actions, PhD thesis, Stanford University. [10] R. M. Gray, Toeplitz and circulant matrices: A review. now publishers (47 Inc, 2006. [I(U,V,i+1 n,y i 1, Z [11] N. Alon, The Shannon capacity of a union, Combinatorica, vol. 18, i+1 n ; no. 3, pp. 301 310, 1998. i [12] R. Ahlswede and I. Csiszár, Common randomness in information I(U,V,i+1 n Y i 1, Z i+1 n ;Y theory and cryptography. i. secret sharing, IEEE Transactions on i] (48 Information Theory, vol. 39, no. 4, pp. 1121 1132, 1993.