On the Capacity Region of the Gaussian Z-channel Nan Liu Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland, College Park, MD 74 nkancy@eng.umd.edu ulukus@eng.umd.edu Abstract We investigate the capacity region of the Gaussian Z-channel with a small crossover link gain, i.e.,. Forthe case of <, we provide an achievable region, and the converse for most of the achievable region. We also derive lower and upper bounds for the part of the region where the capacity boundary is unclear. For the case of =, we determine the capacity region exactly. T R I. INTRODUCTION With recent advances in the theory, technology and applications of wireless communications, the network structures of interest are moving away from traditional networks such as the multiple access channel MAC and the broadcast channel BC which model cellular systems, and towards arbitrary network structures which model ad-hoc wireless networks. One important such network is the interference channel [] [6]. An interference channel is a simple two-transmitter tworeceiver network, where each transmitter has a message for only one of the receivers. A more general network structure is the X-channel [7], where the channel is the same as the interference channel except that both transmitters have messages for both receivers. Unfortunately, most problems in network information theory, including the simple interference channel, have been open for quite a number of years. Recently, [7] has proposed a new multiuser model, called the Z-channel; see Figure. The Z-channel is a special case of the X-channel in that there is only one crossover link and as a consequence, the transmitter that does not have a crossover link has only one message to send. In [7], an achievable region for the Gaussian Z-channel is provided for the case of > +P. In this paper, we focus on the model of the Gaussian Z- channel where the crossover link is weak, more specifically, <. We derive an achievable region and claim that this region is almost equal to the capacity region by proving most of the converse. We also derive some lower and upper bounds on the capacity region. Finally, for the special case of =, we determine the capacity region exactly. II. SYSTEM MODEL The Gaussian Z-channel has two transmitters and two receivers as shown in Figure. The received signals at receivers R and R are given as, Y = X + X + Z Y = X + Z This work was supported by NSF Grants ANI -533 and CCR 3-3; and ARL/CTA Grant DAAD 9---. T Fig.. The Z-channel. where X and X are the signals transmitted by transmitters T and T, and Z,Z are independent Gaussian random variables with zero mean and unit variance. The transmitters T and T are subject to power constraints P and P, respectively. The received signals in and can equivalently be written as, R Y = X + X + Z 3 Y = X + Z 4 since scaling does not affect the capacity region. For the rest of this paper, we will be working with the channel model in 3 and 4. Three independent messages are transmitted in a Z-channel: the message from transmitter T to receiver R, denoted as W, the message from transmitter T to receiver R, denoted as W, and the message from transmitter T to receiver R, denoted as W. The capacity region of the Z-channel is a three dimensional volume, with axes R, R and R corresponding to the rates of messages W, W and W. In this paper, we mainly study the case of <. Reference [7] studied the case of > +P. These two cases correspond to two different kinds of degradedness conditions on the channels from transmitter T to both receivers. In the absence of the link between transmitter T and receiver R, the channels from transmitter T to both receivers constitute a traditional BC []. Given the existence of the link from transmitter T to receiver R, the condition, > +P assumed in [7], corresponds to the case that the signal of transmitter T received at receiver R is a degraded version of the same signal received at receiver R for Gaussian inputs. The condition, <, that we assume in this paper, corresponds to the case that the signal of transmitter T received at receiver R is a degraded version of the same signal at receiver R. The degradedness condition we have here is stronger than the one in [7], in that, it is valid for all
distributions of X. In this paper, we consider only deterministic encoders, which incur no loss in performance [8]. All logarithms are defined with respect to base e..9.8 T U III. ACHIEVABLE REGION Let us define four quantities: c β = + log P 5 βp + c β = log + βp 6 βp + c β = log + βp 7 c β = log + P + βp 8 βp + The following theorem states an achievable region for the Gaussian Z-channel when <. Theorem If <, the following region is achievable in the Gaussian Z-channel: for any β. R c β 9 R c β R c β R + R c β Before providing the proof of Theorem, we show an example of the achievable region in Figure, where P =, P =5and =.5. The boundary of the capacity region is traced as we change β from to, and interpret inequalities in 9- as equalities. Each fixed β determines a pentagon on a plane parallel to the R -R plane as defined by inequalities 9, and, and also a rate R as defined by inequality. Therefore, the achievable region is a concatenation of pentagons of varying sizes along the R axis. Proof Theorem : For simplicity, we will not present probability of error calculations, but rather, we will describe a scheme the transmitters and receivers may use to achieve the region given in 9 to. Fix a β between and, it suffices to show that the two rate triplets: R,R,R = c β,c β c β,c β and R,R,R =c β c β,c β,c β are achievable. This is because, if these two triplets are achievable, then all other points of the region can be achieved by the usual time-sharing technique. First, we will show that R,R,R =c β,c β c β,c β can be achieved. Transmitter T dedicates βp power for transmitting message W using codebook C, and βp power for transmitting message W using codebook C. It transmits the sum of the two codewords. Transmitter T uses all its power P for transmitting message W using codebook C. R.7.6.5.. R. R. S Fig...7.6 V.5 R The achievable region. Receiver R looks at codebook C only, treating everything else as noise, and therefore obtains a rate of R = log + βp P + βp + = c β c β 3 Then, it subtracts the effect of W off, looks at codebook C, treating everything else as noise, and obtains a rate of R = P + log βp + = c β 4 Together, this is a rate of R + R = c β. Receiver R, looks at codebook C only, treating everything else as noise, since R = log + βp P + βp + 5 log + βp 6 βp + it can decode W without error. Subtracting the effect of W off, looking at codebook C, receiver R gets a rate of R = log + βp =c β 7 Thus, rate triplet c β,c β c β,c β is achieved. When both transmitters and receiver R operate in exactly the same way as explained above, and receiver R performs the successive decoding in the reverse order i.e., it decodes W first and then W, the rate triplet R,R,R = c β c β,c β,c β is achieved. We have established the achievability of the region defined by 9-. Next, we will investigate the converse of this achievable region. We will show that, in most places, the achievable region is actually tight, i.e., it is equal to the capacity region of the channel...
IV. THE CONVERSE Theorem The achievable rate triplets R,R,R have to satisfy R c β 8 R c β 9 R + R c β for some β. Referring back to Figure, this theorem states that, of the three surfaces that make up the achievability region, two of them, the surface defined by TRSU and the surface defined by USV, are actually tight. Proof Theorem : We will prove this by using ideas similar to El Gamal s alternative proof [9] to Bergmans proof []. Since there is no cooperation between the two receivers, the capacity region of this channel depends on the joint distribution py,y x,x only through the two marginals py x,x and py x,x [4]. Thus, we will concentrate on the following channel which will yield the same capacity region as our original channel 3-4, Y = X + Y + Z Y = X + Z where Z and Z are Gaussian random variables with zero mean and variance and, respectively. Let rate triplets R,R,R be achievable. Then by Fano s inequality [], there exists an ɛ n such that HW Y n nɛ n 3 HW,W Y n nɛ n 4 and as n, ɛ n. We develop a series of bounds on R, nr = HW 5 = HW Y n +IW ; Y n 6 HW Y n +IW ; Y n W 7 = HW Y n +hy n W hy n W,W 8 = HW Y n +hy n W hz n 9 nɛ n + hy n W n logπe 3 where 7 is obtained from 6 using the independence of messages W and W, 9 is obtained from 8 because we consider deterministic encoders, thus given W and W, we know X n, and therefore the only remaining randomness is in Z n. Finally, 3 follows from 3 and the fact that Z n is an i.i.d. Gaussian sequence with unit variance. Next, we develop a bound for R, nr =HW 3 =HW Y n +IW ; Y n 3 HW,W Y n +IW ; Y n W 33 =HW,W Y n +hy n W hy n W,W 34 nɛ n + hy n W hy n W,W 35 nɛ n + n logπe P + hy n W,W 36 Finally, we develop a bound for R + R, nr + R =HW,W 37 =HW,W Y n +IW,W ; Y n 38 =HW,W Y n +hy n hy n W,W 39 nɛ n + hy n hy n W,W 4 nɛ n + n logπe P + P + hy n W,W 4 where 36 and 4 follow from [, Lemma ]. Consider the following series of inequalities, n logπe =hy n W,W,W 4 hy n W,W 43 hy n W 44 n logπe P + 45 Thus, there exists a β [, ], such that hy n W,W = n logπe βp + 46 From 36, 4 and 46, we see that there exists a β [, ] such that nr nɛ n + nc β 47 nr + R nɛ n + nc β 48 Finally, for R, we argue as follows, hy n W,W =h Xn + Y n + Z n W,W 49 = hy n + Z n W,W 5 = hy n + Z n W 5 where 5 follows because X n is a deterministic function of W, and 5 follows because Y n and Z n are independent of W. Now, let us consider hy n + Z n W. We know that h Z n W =h Z n = n logπe 5 Applying entropy power inequality [, Lemma II], we have hy n + Z n W n logπe e n hy n W + πe 53
Combining 53 with 46 and 5, we have hy n W n logπeβp + 54 Thus, from 3, we have nr nɛ n + c β 55 Since ɛ n as n, using 47, 48 and 55, we obtain the inequalites 8, 9 and, proving the theorem. The converse that is missing is the part that describes R, when R is so small that R + R <c β. This will be addressed in the discussion section next by developing some lower and upper bounds on the capacity region. V. DISCUSSION As stated above, combining Theorems and, we see that the achievable region in Theorem for R,R,R + R is in fact tight. The only unsureness comes from R. As mentioned in [7], the Z-channel includes the MAC, the BC and the Z-interference channel as special cases. By setting β =in the achievable region in Theorem, we get R log + P 56 R log + P 57 R + R log + P + P 58 R = 59 which is exactly the capacity region for the Gaussian MAC with link gains and, and noise variance []. By setting P =in the achievable region in Theorem, we get R log + βp 6 R log + βp 6 βp + R = 6 which is exactly the capacity region for the Gaussian BC with channel gains and, and noise variance []. By setting R =in the capacity region of the Gaussian Z-channel, we should get the capacity region of the Gaussian Z-interference channel [4], which is still an open problem. A. Sum Capacity of the Gaussian Z-channel Similar to the Z-interference channel case, the sum capacity of the Gaussian Z-channel is known for < based on the achievable region of Theorem and the converse theorem, Theorem. The sum capacity is max c β+c β 63 β It can be easily verified that when β =, we attain the maximum and the sum capacity for the Gaussian Z-channel when < is + log P + P + P 64 +P The sum capacity is attained at point U in Figure. R.9.8.7.6.5.... R.7 Fig. 3. All achievable regions and the upper bound in 65. B. Lower and Upper Bounds for the Capacity Region Next, we will derive lower and upper bounds for the capacity region for the portion where a converse is missing. An obvious upper bound for the capacity region is obtained by combining.6 B.5 R R log + P 65 with, and for any β. In Figure 3, the achievable region in Theorem is shown in black and this upper bound is shown in yellow. Two other achievable regions can be derived to close the gap between the lower and upper bounds on the capacity region. Larger Achievable Region : It is clear that the following three points given as triplets of R,R,R are achievable. Point A: log + P,, log + P 66 P + Point B: log + P, log C + P P +., A. 67 Point C: log + P,, P + log + P 68 These three points are shown in Figure 3. Joining the lines between points A and B and points A and C, and the curve connecting points B and C, we can obtain a plane which is achievable by time sharing. It is also worth noting that the line AB is optimal by the same argument as in [4, Theorem ], i.e., line AB is achievable and no point above line AB on the plane of R = log + P is achievable. Larger Achievable Region : Using the technique of succesive decoding [3], we can split X into three parts: X = X + X comm + X priv 69 where X is a function of message W and X comm + X priv together carry message W.LetX comm and X priv be independent. X comm is intended to be decoded by both
receiver R and receiver R, even though receiver R is not interested in decoding any part of message W. X priv is decoded by receiver R only. Receiver R treats X priv as noise. Transmitter T uses power γp for X,powerµγP for X comm, and power µγp for X priv, where γ = γ, µ = µ and γ and µ varyfromto. Receiver R uses decoding order X, then X comm and finally X, and receiver R uses decoding order X, then X comm and finally X priv.letaµ, γ be the set of R, R and R that satisfies the following inequalities: R + log γp P + + γp 7 R comm + log µγp P + + µγp 7 R P + log + µγp 7 R priv log + µγp 73 R =R comm + R priv 74 Receiver R uses decoding order X comm, then X and finally X, and receiver R uses decoding order X comm, then X and finally X priv.letaµ, γ be the set of R, R and R that satisfies the following inequalities: R comm + log µγp P + + µγp 75 + γp R + log γp P + + µγp 76 R log + P + µγp 77 R priv log + µγp 78 R =R comm + R priv 79 Then, an achievable region for the part where the converse is missing is the convex hull of Aµ, γ Aµ, γ 8 µ,γ µ,γ Figure 3 shows Larger Achievable Region and the lines AB and AC defined in Larger Achievable Region in blue. As we can see, there is still a gap between the lower and upper bounds, and additional research is needed to find the exact capacity region. We would like to mention here that using a coding scheme similar to [], we would get an even larger achievable region than Larger Achievable Region. C. The Capacity Region when = Finally, it is worth noting that the Gaussian Z-channel with < has the same capacity region as the channel in Figure 4 where Z and Z are zero-mean Gaussian random variables X / X Fig. 4. Z Y Z The equivalent channel. with variance and, respectively. This can be proved by the argument in [4, Appendix A]. Noting that the two channels have the same capacity is useful, since the capacity region of the channel in Figure 4 might be easier to determine in some cases. For example, for =, Y and Y are statistically equivalent, thus both receiver R and receiver R are able to decode all three messages, W, W and W, similar to a MAC. Thus, the capacity region of the Gaussian Z-channel with =is Y R log + P 8 R + R log + P 8 R + R + R log + P + P 83 VI. CONCLUSION In this paper, we provided an achievable region for the recently proposed Gaussian Z-channel when <. We were able to prove most of the converse for this achievable region. We also provided an upper bound and two larger achievable regions to characterize the capacity region better. We determined the exact capacity region when =. REFERENCES [] T. M. Cover and J. A. Thomas, Elements of Information Theory. Wiley- Interscience, 99. [] E. van der Meulen, Some reflections on the interference channels. Kluwer Academic Publishers, 994, pp. 49 4. [3] A. Carleial, Interference channels, IEEE Trans. Inform. Theory, vol. 4, pp. 6 7, Jan. 978. [4] M. Costa, On the Gaussian interference channel, IEEE Trans. Inform. Theory, vol. 3, pp. 67 65, Sep. 985. [5] G. Kramer, Outer bounds on the capacity of Gaussian interference channels, IEEE Trans. Inform. Theory, vol. 5, pp. 58 586, Mar. 4. [6] I. Sason, On acheivable rate regions for the Gaussian interference channel, IEEE Trans. Inform. Theory, vol. 5, pp. 345 356, Jun. 4. [7] S. Vishwanath, N. Jindal, and A. Goldsmith, The Z channel, in IEEE Globecom, San Francisco, CA, Dec. 3. [8] F. Willems and E. van der Meulen, The discrete memoryless multiple access channel with cribbing encoders, IEEE Trans. Inform. Theory, vol. 3, pp. 33 37, May 985. [9] A. El Gamal, EE478 Multiple User Information Theory, Fall, Lecture notes. [] P. Bergmans, A simple converse for broadcast channels with additive white Gaussian noise, IEEE Trans. Inform. Theory, vol., pp. 79 8, Mar. 974. [] A. El Gamal, The capacity of the physically degraded Gaussian broadcast channel with feedback, IEEE Trans. Inform. Theory, vol. 7, pp. 58 5, Jul. 98. [] K. Marton, A coding theorem for the discrete memoryless broadcast channel, IEEE Trans. Inform. Theory, vol. 5, pp. 36 3, May 979.