Capacity bounds for multiple access-cognitive interference channel

Similar documents
Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels

Multicoding Schemes for Interference Channels

An Outer Bound for the Gaussian. Interference channel with a relay.

Information Theory. Lecture 10. Network Information Theory (CT15); a focus on channel capacity results

Interference Channels with Source Cooperation

Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets

Lecture 10: Broadcast Channel and Superposition Coding

820 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 58, NO. 2, FEBRUARY Stefano Rini, Daniela Tuninetti, and Natasha Devroye

On the Capacity of the Interference Channel with a Relay

On Multiple User Channels with State Information at the Transmitters

The Capacity Region of the Gaussian Cognitive Radio Channels at High SNR

5958 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 12, DECEMBER 2010

Simultaneous Nonunique Decoding Is Rate-Optimal

Relay Networks With Delays

Information Theory for Wireless Communications. Lecture 10 Discrete Memoryless Multiple Access Channel (DM-MAC): The Converse Theorem

Cognitive Multiple Access Networks

Efficient Use of Joint Source-Destination Cooperation in the Gaussian Multiple Access Channel

Lecture 1: The Multiple Access Channel. Copyright G. Caire 12

On the K-user Cognitive Interference Channel with Cumulative Message Sharing Sum-Capacity

The Gallager Converse

On the Capacity Region of the Gaussian Z-channel

Interference Channel aided by an Infrastructure Relay

Capacity of a Class of Cognitive Radio Channels: Interference Channels with Degraded Message Sets

EE 4TM4: Digital Communications II. Channel Capacity

Lecture 5: Channel Capacity. Copyright G. Caire (Sample Lectures) 122

On the Capacity of Interference Channels with Degraded Message sets

Feedback Capacity of the Gaussian Interference Channel to Within Bits: the Symmetric Case

arxiv: v2 [cs.it] 28 May 2017

On Gaussian MIMO Broadcast Channels with Common and Private Messages

An Achievable Rate Region for the 3-User-Pair Deterministic Interference Channel

Outer Bounds on the Secrecy Capacity Region of the 2-user Z Interference Channel With Unidirectional Transmitter Cooperation

Source-Channel Coding Theorems for the Multiple-Access Relay Channel

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

Secrecy in the 2-User Symmetric Interference Channel with Transmitter Cooperation: Deterministic View

ELEC546 Review of Information Theory

Clean relaying aided cognitive radio under the coexistence constraint

Error Exponent Region for Gaussian Broadcast Channels

arxiv: v1 [cs.it] 4 Jun 2018

Random Access: An Information-Theoretic Perspective

LECTURE 13. Last time: Lecture outline

On the Capacity and Degrees of Freedom Regions of MIMO Interference Channels with Limited Receiver Cooperation

Generalized Writing on Dirty Paper

Approximately achieving the feedback interference channel capacity with point-to-point codes

Source and Channel Coding for Correlated Sources Over Multiuser Channels

Classical codes for quantum broadcast channels. arxiv: Ivan Savov and Mark M. Wilde School of Computer Science, McGill University

Joint Source-Channel Coding for the Multiple-Access Relay Channel

The Capacity Region of the Cognitive Z-interference Channel with One Noiseless Component

Secret Key Agreement Using Conferencing in State- Dependent Multiple Access Channels with An Eavesdropper

Primary Rate-Splitting Achieves Capacity for the Gaussian Cognitive Interference Channel

Capacity Region of Reversely Degraded Gaussian MIMO Broadcast Channel

Shannon s noisy-channel theorem

Approximate Capacity of Fast Fading Interference Channels with no CSIT

A Comparison of Two Achievable Rate Regions for the Interference Channel

Bounds and Capacity Results for the Cognitive Z-interference Channel

Research Article Multiaccess Channels with State Known to Some Encoders and Independent Messages

On Capacity Under Received-Signal Constraints

K User Interference Channel with Backhaul

A Comparison of Superposition Coding Schemes

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

List of Figures. Acknowledgements. Abstract 1. 1 Introduction 2. 2 Preliminaries Superposition Coding Block Markov Encoding...

Research Article Lower Bounds on the Capacity of the Relay Channel with States at the Source

A Half-Duplex Cooperative Scheme with Partial Decode-Forward Relaying

On the Capacity of the Two-Hop Half-Duplex Relay Channel

The Robustness of Dirty Paper Coding and The Binary Dirty Multiple Access Channel with Common Interference

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions

Capacity of channel with energy harvesting transmitter

Solutions to Homework Set #2 Broadcast channel, degraded message set, Csiszar Sum Equality

A Proof of the Converse for the Capacity of Gaussian MIMO Broadcast Channels

Lecture 2. Capacity of the Gaussian channel

II. THE TWO-WAY TWO-RELAY CHANNEL

Secret Key Agreement Using Asymmetry in Channel State Knowledge

AN INTRODUCTION TO SECRECY CAPACITY. 1. Overview

The Capacity Region of the Gaussian MIMO Broadcast Channel

On Network Interference Management

Capacity Bounds for. the Gaussian Interference Channel

Cognitive Radio: An Information-Theoretic Perspective

Achievable Rates and Outer Bound for the Half-Duplex MAC with Generalized Feedback

Variable Length Codes for Degraded Broadcast Channels

EE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15

On the Duality between Multiple-Access Codes and Computation Codes

On Achievability for Downlink Cloud Radio Access Networks with Base Station Cooperation

Sum Capacity of Gaussian Vector Broadcast Channels

THE multiple-access relay channel (MARC) is a multiuser

UCSD ECE 255C Handout #12 Prof. Young-Han Kim Tuesday, February 28, Solutions to Take-Home Midterm (Prepared by Pinar Sen)

The Poisson Channel with Side Information

IET Commun., 2009, Vol. 3, Iss. 4, pp doi: /iet-com & The Institution of Engineering and Technology 2009

On Scalable Coding in the Presence of Decoder Side Information

ProblemsWeCanSolveWithaHelper

Lecture 4 Noisy Channel Coding

National University of Singapore Department of Electrical & Computer Engineering. Examination for

On the Secrecy Capacity of Fading Channels

1174 IET Commun., 2010, Vol. 4, Iss. 10, pp

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information

i.i.d. Mixed Inputs and Treating Interference as Noise are gdof Optimal for the Symmetric Gaussian Two-user Interference Channel

Polar codes for the m-user MAC and matroids

ECE Information theory Final (Fall 2008)

UCSD ECE 255C Handout #14 Prof. Young-Han Kim Thursday, March 9, Solutions to Homework Set #5

The Capacity of the Semi-Deterministic Cognitive Interference Channel and its Application to Constant Gap Results for the Gaussian Channel

On Compound Channels With Side Information at the Transmitter

Two Applications of the Gaussian Poincaré Inequality in the Shannon Theory

Transcription:

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 RESEARCH Open Access Capacity bounds for multiple access-cognitive interference channel Mahtab Mirmohseni *, Bahareh Akhbari and Mohammad Reza Aref Abstract Motivated by the uplink scenario in cellular cognitive radio, this study considers a communication network in which a point-to-point channel with a cognitive transmitter and a Multiple Access Channel (MAC) with common information share the same medium and interfere with each other. A Multiple Access-Cognitive Interference Channel (MA-CIFC) is proposed with three transmitters and two receivers, and its capacity region in different interference regimes is investigated. First, the inner bounds on the capacity region for the general discrete memoryless case are derived. Next, an outer bound on the capacity region for full parameter regime is provided. Using the derived inner and outer bounds, the capacity region for a class of degraded MA-CIFC is characterized. Two sets of strong interference conditions are also derived under which the capacity regions are established. Then, an investigation of the Gaussian case is presented, and the capacity regions are derived in the weak and strong interference regimes. Some numerical examples are also provided. Keywords: Cognitive interference channel, Multiple access channel, Strong Interference, Weak interference, Capacity region. Introduction Interference avoidance techniques have traditionally been used in wireless networks wherein multiple source-destination pairs share the same medium. However, the broadcasting nature of wireless networks may enable cooperation among entities, which ensures higher rates with more reliable communication. On the other hand, due to the increasing number of wireless systems, spectrum resources have become scarce and expensive. The exponentially growing demand for wireless services along with the rapid advancements in wireless technology has lead to cognitive radio technology which aims to overcome the spectrum inefficiency problem by developing communication systems that have the capability to sense the environment and adapt to it []. In overlay cognitive networks, the cognitive user can transmit simultaneously with the non-cognitive users and compensate for the interference by cooperation in sending, i.e., relaying, the non-cognitive users messages []. From an information theoretic point of view, Cognitive Interference Channel (CIFC) was first introduced in [] * Correspondence: mirmohseni@ee.sharif.edu Department of Electrical Engineering, Information Systems and Security Lab (ISSL) Sharif University of Technology, Tehran, Iran to model an overlay cognitive radio and refers to a twouser Interference Channel (IFC) in which the cognitive user (secondary user) has the ability to obtain the message being transmitted by the other user (primary user), either in a non-causal or in a causal manner. An achievable rate region for the non-causal CIFC was derived in [], by combining the Gel fand-pinsker (GP) binning [3] with a well-known simultaneous superposition coding scheme (rate splitting) applied to IFC [4]. For the noncausal CIFC, where the cognitive user has non-causal full or partial knowledge of the primary user s transmitted message several achievable rate regions and capacity results in some special cases have been established [5-4]. More recently a three-user cognitive radio network with one primary user and two cognitive users is studied in [5,6], where an achievable rate region is derivedforthissetupbased on rate splitting and GP binning. In the interference avoidance-based systems, i.e., when the communication medium is interference-free, uplink transmission is modeled with a Multiple Access Channel (MAC) whose capacity region has been fully characterized for independent transmitters [7,8] as well as for the transmitters with common information [9]. Recently, Mirmohseni et al; licensee Springer. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/.), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page of 8 taking the effects of interference into account in the uplink scenario, a MAC and an IFC have been merged into one setup by adding one more transmit-receive pair to the communication medium of a two-user MAC [,], where the channel inputs at the transmitters are independent and there is no cognition or cooperation. In this paper, we introduce Multiple Access-Cognitive Interference Channel (MA-CIFC) by providing the transmitter of the point-to-point channel with cognition capabilities in the uplink with interference model. Moreover, transmitters of MAC have common information that enables cooperation among them. As shown in Figure, the proposed channe consists of three transmitters and two receivers: two-user MAC with common information as the primary network and a point-to-point channel with a cognitive transmitter that knows the message being sent by all of the transmitters in a non-causa manner. A physical example of this channel is the coexistence of cognitive users with the licensed primary users in a cellular or satellite uplink transmission, where the cognitive radios by their abilities exploit side information about the environment to maintain or improve the communication of primary users while also achieving some spectrum resources for their own communication. In this scenario, the primary non-cognitive users can be oblivious to the or aware of the cognitive users []. When the non-cognitive user is oblivious to the cognitive user s presence, its receiver s decoding process is independent of the interference caused by the cognitive user s transmission. In fact, the primary receiver treats interference as noise. However, in the aware cognitive user s scenario, the decoding process at the primary receiver can be adapted to improve its own rate. For example, the primary receiver can decode the cognitive user s message and cancel the interference when the interfering signal is strong enough. If the multi-antenna capability is available at the primary receiver, it can also m m mˆ, mˆ ˆ, m, m, m m, m, m, m 3 Cognitive user Figure Graphic representation for MA-CIFC. ˆm 3 reduce or increase the interfering signal by beam-steering, which results in the occurrence of the weak or strong interference regimes []. To analyze the capacity region of MA-CIFC, we first derive three inner bounds on the capacity region (achievable rate regions). The first two bounds assume an oblivious primary receiver, which does not decode the cognitive user s message but treats it as noise. Two different coding schemes are proposed based on the superposition coding, the GP binning and the method of [6] in defining auxiliary Random Variables (RVs). Later, we show that these strategies are optimal for a degraded MA-CIFC and also in the Gaussian weak interference regime. In the third achievability scheme, we consider an aware primary receiver and obtain an inner bound on the capacity region based on using superposition coding in the encoding part and allowing both receivers to decode all messages with simultaneous joint decoding in the decoding part. This strategy is capacity-achieving in the strong interference regime. Next, we provide a general outer bound on the capacity region and derive conditions under which the first achievability scheme achieves capacity for the degraded MA-CIFC. We continue the capacity results by the derivation of two sets of strong interference conditions, under which the third inner bound achieves capacity. Further, we compare these two sets of conditions and identify the weaker set. We also extend the strong interference results to a network with k primary users. Moreover, we consider the Gaussian case and find capacity results for the Gaussian MA-CIFC in both the weak and strong interference regimes. We use the second derived inner bound to show that the capacityachieving scheme in weak interference consists of Dirty Paper Coding (DPC) [] at the cognitive transmitter and treating interference as noise at both receivers. We also provide some numerical examples. The rest of the paper is organized as follows. Section introduces MA-CIFC model and the notations. Three inner bounds and an outer bound on the capacity region are derived in Section 3 and Section 4, respectively, for the discrete memoryless MA-CIFC. Sections 5 presents the capacity results for the discrete memoryless MA- CIFC in three special cases. In Section 6, the Gaussian MA-CIFC is investigated. Finally, Section 7 concludes the paper.. Channel models and preliminaries Throughout the paper, upper case letters (e.g. X) are used to denote RVs and lower case letters (e.g. x) show their realizations. The probability mass function (p.m.f) of a RV X with alphabet set X is denoted by p X (x), where subscript X is occasionally omitted. A n ε (X, Y) specifies the set of -strongly, jointly typical sequences of

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page 3 of 8 length n. ThenotationX j indicates a sequence of RVs i (X i, X i+,..., X j ), where X j is used instead of X j, for brevity. N (, σ ) denotes a zero mean normal distribution with variance s. Consider the MA-CIFC in Figure, which is denoted by (X X X 3, p(y n, yn 3 xn, xn, xn 3 ), Y Y 3 ), where X X, X X and X 3 X 3 are channel inputs at Transmitter (Tx), Transmitter (Tx) and Transmitter 3 (Tx3), respectively; Y Y and Y 3 Y 3 are channel outputs at Receiver (Rx) and Receiver 3 (Rx3), respectively; and p(y n, yn 3 xn, xn, xn 3 ) is the channel transition probability distribution. In n channel uses, each Txj desires to send a message pair (m, m j )torxwherej Î {,}, and Tx3 desires to send a message m 3 to Rx3. Definition : A ( nr, nr, nr, nr 3, n) code for MA- CIFC consists of (i) four independent message sets M j {,..., nr j}, where j Î {,,, 3}; (ii) two encoding functions at the primary transmitters, f : M M X n at Tx and f : M M X n at Tx; (iii) an encoding function at the cognitive transmitter, f 3 : M M M M 3 X3 n ; and (iv) two decoding functions, g : Y n M M M at Rx and g 3 : Y3 n M 3 at Rx3. We assume that the channel is memoryless. Thus, the channel transition probability distribution is given by n p(y n, yn 3 xn, xn, xn 3 ) p(y,i, y 3,i x,i, x,i, x 3,i ). () The probability of error for this code is defined as P e n (R + R + R + R 3 ) m,m,m,m 3 p[{g 3 (Y n 3 ) m 3} {g (Y n ) (m, m, m )} (m, m, m, m 3 )sent]. Definition : A rate quadruple (R, R, R, R 3 )is achievable if there exists a sequence of ( nr, nr, nr, nr s, n) codes with P e asn. The capacity region C, is the closure of the set of all achievable rates. 3. Inner bounds on the capacity region of discrete memoryless MA-CIFC Now,wederivethreeachievablerateregionsforthe general setup. Theorems and assume an oblivious primary receiver (Rx), which does not decode the cognitive user s message (m 3 ) and treats it as noise. The decoding procedure at the cognitive receiver (Rx3) differs in these schemes. In Theorem, the cognitive receiver (Rx3) decodes the primary messages (m, m, m ), and all the transmitters use superposition coding. However, in Theorem, the cognitive receiver (Rx3) also treats the interference from the primary messages (m, m, m ) as noise, while the cognitive transmitter (Tx3) uses GP binning to precode its message for interference cancelation at Rx3. We also utilize the method of [6] in defining auxiliary RVs, which helps us to achieve the outer bound in special cases. In fact, we achieve the outer bound of Theorem 4 using the region of Theorem m, m Primary network X n Transmitter (Tx) Receiver (Rx) mˆ, mˆ, mˆ n n m, m Transmitter X Y (Tx) p y, y 3 x, x, x3 X n m, m, m Transmitter 3 3 ˆm 3 Receiver 3, m 3 (Tx3) (Rx3) Cognitive user Y n 3 Figure Multiple Access-Cognitive Interference Channel (MA-CIFC).

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page 4 of 8 for a class of degraded MA-CIFC in Section 5. The region of Theorem is used in Section 6 to derive the capacity region in the weak interference regime. In the scheme of Theorem 3, we consider an aware primary receiver (Rx) which decodes the cognitive user s message (m 3 ). The cognitive receiver (Rx3) also decodes the primary messages (m, m, m ). Therefore, this region is obtained based on using superposition coding in the encoding part and by allowing both receivers to decode all messages with simultaneous joint decoding in the decoding part. In Section 5, we show that this strategy is capacity-achieving in the strong interference regime. Proofs are provided in Appendix A. Theorem : The union of rate regions given by R 3 I(X 3 ; Y 3 T, U, X, V, X ) () R I(U, X ; Y T, V, X ) (3) R I(V, X ; Y T, U, X ) (4) R + R I(U, X, V, X ; Y T) (5) R + R + R I(T, U, X, V, X ; Y ) (6) R + R 3 I(U, X, X 3 ; Y 3 T, V, X ) (7) R + R 3 I(V, X, X 3 ; Y 3 T, U, X ) (8) R + R + R 3 I(U, X, V, X, X 3 ; Y 3 T) (9) R + R + R + R 3 I(T, U, X, V, X, X 3 ; Y 3 ) () is achievable for MA-CIFC, where the union is over all p.m.fs that factor as p(t)p(u, x t)p(v, x t)p(x 3 t, u, x, v, x ). () Theorem : The union of rate regions given by (3)-(6) and R 3 I(W; Y 3 ) I(W; T, U, X, V, X ) () is achievable for MA-CIFC, where the union is over all p.m.fs that factor as p(t)p(u, x t)p(v, x t)p(w, x 3 t, u, x, v, x ). (3) Theorem 3: The union of rate regions given by R 3 I(X 3 ; Y 3 X, X, T) (4) R + R 3 min{i(x, X 3 ; Y X, T), I(X, X 3 ; Y 3 X, T)} R + R 3 min{i(x, X 3 ; Y X, T), I(X, X 3 ; Y 3 X, T)} R + R + R + R 3 min{i(x, X, X 3 ; Y ), I(X, X, X 3 ; Y 3 )} (5) (6) (7) is achievable for MA-CIFC, where the union is over all p.m.fs that factor as p(t)p(x t)p(x t)p(x 3 x, x, t). (8) Remark : We utilize the region of Theorem in Section 5 to achieve capacity results for a class of degraded MA-CIFC, and the region of Theorem to derive the results for the Gaussian case in Section 6. The region of Theorem 3 is also used to characterize the capacity region under strong interference conditions in Section 5. 4. An outer bound on the capacity region of discrete memoryless MA-CIFC Here, we derive a general outer bound on the capacity region of MA-CIFC which is used to obtain the capacity region for a class of the degraded MA-CIFC in Section 5 and also to find capacity results for the Gaussian MA- CIFC in the weak interference regime in Section 6. Let R o denote the union of all rate quadruples (R, R, R, R 3 ) satisfying (3)-(6) and R 3 I(X 3 ; Y 3, Y T, U, X, V, X ), (9) where the union is over all p.m.fs that factor as (). Theorem 4: The capacity region of MA-CIFC satisfies C R o Proof: Consider a ( nr, nr, nr, nr 3, n) code with the average error probability of Pe n. Define the following RVs for i,..., n: T i (M, Y i ) () U i (M, M, Y i )(M, T i ) () V i (M, M, Y i )(M, T i ) () Considering the encoding functions f and f, defined in Definition, and the above definitions for auxiliary

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page 5 of 8 RVs, we remark that (X,i, U i ) T i (X,i, V i ) forms a Markov chain. Thus, these choices of auxiliary RVs satisfy the p.m.f () of Theorem 4. Now using Fano s inequality [3], we derive the bounds in Theorem 4. For the first bound, we have: nr 3 H(M 3 ) H(M 3 M, M, M ) I(M 3 ; Y3 n M, M, M )+H(M 3 Y3 n, M, M, M ) (3) I(M 3 ; Y3 n M, M, M )+nδ 3n where follows since messages are independent and holds due to Fano s inequality and the fact that conditioning does not increase entropy. Hence, nr 3 nδ 3n I(M 3 ; Y n 3 M, M, M ) I(M 3, X3 n ; Yn 3, Yn M, M, M, X n, Xn ) ( ) I M 3, X3 n ; Y 3,i, Y,i M, M, M, X n, Xn, Yi 3, Y i (c) (d) ( H Y 3,i, Y,i M, M, M, X,i, X,i, Y i H(Y 3,i, Y,i M, M, M, Y i, X,i, X,i, X 3,i ) I(X 3,i ; Y 3,i, Y,i T i, U i, X,i, V i, X,i ) ) (4) where is due to the encoding functions f, f and f 3, defined in Definition, and the non-negativity of mutual information, is obtained from the chain rule, (c) follows from the memoryless property of the channel and the fact that conditioning does not increase entropy, and (d) is obtained from ()-(). Now, applying Fano s inequality and the independence of the messages, we can bound R as: nr nδ n I(M ; Y n M, M ) I(M, X,i ; Y,i M, M, X,i, Y i ) I(M, X,i, M, Y i ; Y,i M, M, X,i, Y i ) I(U i, X,i ; Y,i T i, V i, X,i ), (5) where follows from the chain rule and the encoding functions f and f, and from ()-(). Similarly, we can show that nr nδ n I(V i, X,i ; Y,i T i, U i, X,i ). (6) Next, based on similar arguments, we bound R + R as n(r + R ) n(δ n + δ n ) I(M, M ; Y n M ) I(M, X,i, M, X,i ; Y,i M, Y i ) I(M, X,i, M, X,i, Y i ; Y,i M, Y i ) I(U i, X,i, V i, X,i ; Y,i T i ). The last sum-rate bound can be derived as follows: n(r + R + R ) n(δ n + δ n + δ n ) I(M, M, M ; Y n ) I(M, M, X,i, M, X,i ; Y,i Y i ) H(Y,i ) H(Y,i M, M, X,i, M, X,i, Y i ) I(T i, U i, X,i, V i, X,i ; Y,i ) (7) (8) where follows since conditioning does not increase entropy. Using the standard time-sharing argument for (4)-(8) completes the proof. 5. Capacity results for discrete memoryless MA- CIFC In this section, we characterize the capacity region of MA-CIFC under specific conditions. First, we consider a class of degraded MA-CIFC and derive conditions under which the inner bound in Theorem achieves the outer bound of Theorem 4. Next, we investigate the strong interference regime by deriving two sets of strong interference conditions under which the region of Theorem 3 achieves capacity. We also compare these two sets of conditions and identify the weaker set. Finally, we extend the strong interference results to a network with k primary users. A. Degraded MA-CIFC Now, we characterize the capacity region for a class of MA-CIFC with a degraded primary receiver. We define MA-CIFC with a degraded primary receiver as a MA- CIFC where Y and X 3 are independent given Y 3, X, X. More precisely, the following Markov chain holds: X 3 X, X Y 3 X, X Y X, X, (9) or equivalently, X 3 (X, X, Y 3 ) Y forms a Markov chain. This means that the primary receiver (Rx) observes a degraded or noisier version of the cognitive user s signal (Tx3) compared with the cognitive receiver (Rx3).

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page 6 of 8 Assume that the following conditions are satisfied for MA-CIFC over all p.m.fs that factor as (): I(U, X ; Y T, V, X ) I(U, X ; Y 3 T, V, X ) (3) I(V, X ; Y T, U, X ) I(V, X ; Y 3 T, U, X ) (3) I(U, X, V, X ; Y T) I(U, X, V, X ; Y 3 T) (3) I(T, U, X, V, X ; Y ) I(T, U, X, V, X ; Y 3 ) (33) Under these conditions, the cognitive receiver (Rx3) can decode the messages of the primary users with no rate penalty. If MA-CIFC with a degraded primary receiver satisfies conditions (3)-(33), the region of Theorem coincideswithr o and achieves capacity, as stated in the following theorem. Theorem 5: ThecapacityregionofMA-CIFCwitha degraded primary receiver, defined in (9), satisfying (3)-(33) is given by the union of rate regions satisfying ()-(6) over all joint p.m.fs (). Remark : The messages of the primary users (m, m, m ) can be decoded at Rx3 under conditions (3)-(33). Therefore, Rx3-Tx3 achieves the rate in (). Moreover, we can see that due to the degradedness condition in (9), treating interference as noise at the primary receiver (Rx) achieves capacity. We show in Section 6 that, in the Gaussian case the capacity is achieved by using the region of Theorem based on DPC (or GP binning), where the cognitive receiver (Rx3) does not decode the primary messages and conditions (3)-(33) are not necessary. Proof: Achievability: The proof follows from the region of Theorem. Using the condition in (3), the sum of the bounds in () and (3) makes the bound in (7) redundant. Similarly, conditions (3)-(33), along with the bound in (), make the bounds in (8)-() redundant and the region reduces to ()-(6). Converse: To prove the converse part, we evaluate R o of Theorem 4 with the degradedness condition in (9). Itisnotedthatthep.m.fofTheorem5isthesameas the one for R o. Moreover, the bounds in (3)-(6) are equal for both regions. Hence, it is only necessary to show the bound in (). Considering (9), we obtain: R 3 I(X 3 ; Y 3, Y T, U, X, V, X ) I(X 3 ; Y 3 T, U, X, V, X )+I(X 3 ; Y T, U, X, V, X, Y 3 ) I(X 3 ; Y 3 T, U, X, V, X ) where is obtained by applying the degradedness condition in (9). This completes the proof. B. Strong interference regime Now, we derive two sets of strong interference conditions under which the region of Theorem 3 achieves capacity. First, assume that the following set of strong interference conditions, referred to as Set, holds for all p.m.fs that factor as (8): I(X 3 ; Y 3 X, X, T) I(X 3 ; Y X, X, T) (34) I(X, X 3 ; Y X, T) I(X, X 3 ; Y 3 X, T) (35) I(X, X 3 ; Y X, T) I(X, X 3 ; Y 3 X, T) (36) I(X, X, X 3 ; Y ) I(X, X, X 3 ; Y 3 ). (37) In fact, under these conditions, interfering signals at the receivers are strong enough that all messages can be decoded by both receivers. Condition (34) implies that the cognitive user s message (m 3 )canbedecodedat Rx, while conditions (35)-(37) guarantee the decoding of the primary messages (m, m, m )alongwithm 3 at Rx3 in a MAC fashion. Theorem 6: The capacity region of MA-CIFC satisfying (34)-(37) is given by: C str { (R, R, R, R 3 ): p(t)p(x t)p(x t)p(x 3 x,x,t) R, R, R, R 3 R 3 I(X 3 ; Y 3 X, X, T) (38) R + R 3 I(X, X 3 ; Y X, T) (39) R + R 3 I(X, X 3 ; Y X, T) (4) R + R + R + R 3 I(X, X, X 3 ; Y ) }. (4) Remark 3: The message of the cognitive user (m 3 )can be decoded at Rx, under condition (34) and (m, m, m ) can be decoded at Rx3 under conditions (35)-(37). Hence, the bound in (38) gives the capacity of a pointto-point channel with message m 3 with side-information X, X at the receiver. Moreover, (38)-(4) with condition (34) give the capacity region for a three-user MAC with common information where R and R are the common rates, R 3 is the private rate for Tx3, and the private rates for Tx and Tx are zero. Remark 4: If we omit Tx, i.e., X, and Tx has no message to transmit, i.e., R, the model reduces to a CIFC, and C str coincides with the capacity region of the strong interference channel with unidirectional cooperation (or CIFC), which was characterized in [8, Theorem 5]. It is noted that in this case, the common message can be ignored, i.e., T and R. Proof: Achievability: Considering (35)-(37), the proof follows from Theorem 3. Converse: Consider a ( nr, nr, nr, nr 3, n) code with an average error probability of Pe n. Definethe following RV for i,..., n: T n M (4)

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page 7 of 8 It is noted that due to the encoding functions f, f and f 3, defined in Definition, the independence of messages, and the above definitions for T n,rvssatisfy thep.m.f(8)oftheorem6.first,weprovideauseful lemma which we need in the proof of the converse part. Lemma : If (34) holds for all distributions that factor as (8), then I(X n 3 ; Yn 3 Xn, Xn, Tn, U) I(X n 3 ; Yn Xn, Xn, Tn, U). (43) Proof: The proof relies on the results in [4, Proposition ] and [5, Lemma]. By redefining X X 3, Y Y 3, X (X, X, T) in [8, Lemma 5], the proof follows. Now, using Fano s inequality [3], we derive the bounds in Theorem 6. Using (3) provides: nr 3 nδ 3n I(M 3 ; Y n 3 M, M, M ) I(M 3, X n 3 ; Yn 3 Tn, M, M, X n, Xn ) I(X3 n ; Yn 3 Tn, X n, Xn ) (c) (d) I(X n 3 ; Y 3,i T n, X n, Xn, Yi 3 ) I(X 3,i ; Y 3,i X,i, X,i, T i ) (44) where is due to (4) and the encoding functions f, f and f 3, defined in Definition, follows from two facts; conditioning does not increase entropy and (M, M, M 3 ) (X n, Xn, Xn 3 ) Yn 3 forms a Markov chain, (c) is obtained from the chain rule, and (d) follows from the memoryless property of the channel and the fact that conditioning does not increase entropy. Now, applying Fano s inequality and the independence of the messages, we can bound R + R 3 as n(r + R 3 ) n(δ n + δ 3n ) I(M ; Y n M, M )+I(M 3 ; Y3 n M, M, M ) I(M, X n ; Yn M, M, X n ) +I(M 3, X3 n ; Yn 3 M, M, M, X n, Xn ) I(M, X n ; Yn Tn, M, X n ) +I(X3 n ; Yn 3 Tn, M, M, X n, Xn ) (c) I(M, X n ; Yn Tn, M, X n ) +I(X3 n ; Yn Tn, M, M, X n, Xn ) I(M, X n, Xn 3 ; Yn Tn, M, X n ) (d) (e) I(M, X n, Xn 3 ; Y,i T n, M, X n, Yi ) I(X,i, X 3,i ; Y,i X,i, T i ) (45) where follows from encoding functions f, f and f 3, follows from (4) and the fact that M 3 (X n, Xn, Xn 3 ) Yn 3 forms a Markov chain, (c) is obtained from (43), (d) follows from the chain rule, and (e) follows from the memoryless property of the channel and the fact that conditioning does not increase entropy. Applying similar steps, we can show that, n(r + R 3 ) n(δ n + δ 3n ) I(X,i, X 3,i ; Y,i X,i, T ). (46) Finally, the sum-rate bound can be obtained as n(r + R + R + R 3 ) n(δ n + δ n + δ n + δ 3n ) I(M, M, M ; Y n )+I(M 3; Y3 n M, M, M ) I(M, M, M, X n, Xn ; Yn ) +I(M 3, X3 n ; Yn 3 M, M, M, X n, Xn ) I(T n, M, M, X n, Xn ; Yn ) +I(X3 n ; Yn Tn, M, M, X n, Xn ) I(T n, M, M, X n, Xn, Xn 3 ; Yn ) I(T n, X n, Xn, Xn 3 ; Yn ) I(X,i, X,i, X 3,i ; Y,i ) (c) (47) where follows from steps -(c) in (45), is due to the fact that (M, M ) (X n, Xn, Xn 3 ) Yn forms a Markov chain, and (c) follows from the memoryless property of the channel and the fact that conditioning does not increase entropy. Using a standard time-sharing argument for (44)-(47) completes the proof. Next, we derive the second set of strong interference conditions, called Set, under which the region of Theorem 3 is the capacity region. For all p.m.fs that factor as (8), Set includes (34) and the following conditions: I(X ; Y X, T) I(X ; Y 3 X, T) (48) I(X ; Y X, T) I(X ; Y 3 X, T) (49) I(X, X ; Y ) I(X, X ; Y 3 ). (5) Remark 5: Similar to the condition Set, under these conditions interfering signals at the receivers are strong enough that all messages can be decoded by both receivers. The first condition in (34) is equal in the two sets under which the cognitive user s message(m 3 ) can be decoded at Rx. However, conditions (48)-(5) imply that the primary messages (m, m, m ) can be decoded at Rx3 in a MAC fashion, while in Set, they can be decoded along with m 3.

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page 8 of 8 Theorem 7: The capacity region of MA-CIFC, satisfying (34) and (48)-(5), referred to as C str, is given by the union of rate regions satisfying (4)-(7) over all p.m.fs that factor as (8). Proof: See Appendix B. Remark 6: Similar to Remark 4, by omitting Tx (T X, R R ), themodelreducestoa CIFC. Moreover, C and Set reduce to the capacity region and strong interference conditions which have been derived in [3] for non-causal CIFC. Remark 7 (Comparison of two sets of conditions): In the strong interference conditions of Set, the first condition in (34) is used in the converse part, while (35)- (37) are used to reduce the inner bound to C str.how- ever, all the conditions of Set are utilized to prove the converse part. Now, we compare the conditions in these two sets. We can write (35) as I(X ; Y X, T)+[I(X 3 ; Y X, X, T) I(X 3 ; Y 3 X, X, T)] }{{} I diff I(X ; Y 3 X, T). Considering (34), it can be seen that I diff. Hence, condition (35) implies condition (48), but not vice versa. Similar conclusions can be drawn for other conditions of these two sets. Therefore, Set implies Set, andthe conditions of Set are weaker compared to those of Set. C. Multiple access-cognitive interference network (MA- CIFN) Now, we extend the result of Theorem 6 to a network with k + transmitters and two receivers; a k-user MAC as a primary network and a point-to-point channel with a cognitive transmitter. We call it Multiple Access-Cognitive Interference Network (MA-CIFN). Consider MA-CIFN in Figure 3, denoted by (X X X k X k+, p(y n, yn k+ xn, xn, xn k, xn k+ ), Y Y k+ ), where X j X j is the channel input at Transmitter j (Txj), for j {,..., k +}; Y Y and Y k+ Y k+ are channel outputs at the primary and cognitive receivers, respectively, and p(y n, yn k+ nn, xn,..., xn k, xn k+ ) is the channel transition probability distribution. In n channel uses, each Txj desires to send a message pair m j to the primary receiver where j Î {,..., k}, and Txk + desires to send a message m k+ to the cognitive receiver. We ignore the common information for brevity. Definitions and can be simply extended to the MA-CIFN. Therefore, we state the result on the capacity region under strong interference conditions. Corollary : The capacity region of the MA-CIFN, satisfying I(X k+ ; Y k+ X([ : k])) I(X k+ ; Y X([ : k])) (5) I(X k+, X(S); Y X(S c )) I(X k+, X(S); Y k+ X(S c )) (5) for all S [: k] andforeveryp(x )p(x )...p(x k ) p(x k+ x,x,...,x k )p(y,y k+ x,x,...,x k,x k+ ), is given by { Cnet str p(x )p(x )...p(x k)p(x k+ x,x,...,x k) (R, R,..., R k, R k+ ):R, R,..., R k, R k+ R k+ I(X k+ ; Y k+ X([ : k])) (53) R k+ + j S R j I(X k+, X(S); Y X(S c ))} (54) for all S [: k], where X(S) istheorderedvectorof X j, j Î S, and S c denotes the complement of the set S. m m X X Y mˆ ˆ,, mk m k X k m m,,, mk, mk X k Y k mˆ k Cognitive user Figure 3 Graphic representation for the MA-CIFN.

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page 9 of 8 Proof: Following the same lines as the proof of Theorem 6, the proof is straightforward. Therefore, it is omitted for the sake of brevity. Remark 8: Under condition (5), the message of the cognitive user (m k+ ) can be decoded at the primary receiver (Y ). Also, the cognitive receiver (Y k+ ), under condition (5), can decode m j ; j Î {,...,k} inamac fashion. Therefore, the bound in (53) gives the capacity of a point-to-point channel with message m k+ with side-information X j ; j Î {,..., k} at the cognitive receiver. Moreover, (53) and (54) with condition (5), give the capacity region for a k+-user MAC with common information at the primary receiver. 6. Gaussian MA-CIFC In this section, we consider the Gaussian MA-CIFC and characterize capacity results for the Gaussian case in the weak and strong interference regimes. For simplicity, we assume that Tx and Tx have no common information. This means that, R andm. to investigate these regions. The Gaussian MA-CIFC, as depicted in Figure 4, at time i,...,n can be mathematically modeled as Y,i X,i + X,i + h 3 X 3,i + Z,i (55) Y 3,i h 3 X,i + h 3 X,i + X 3,i + Z 3,i (56) where h 3, h 3,andh 3 are known channel gains. X,i, X,i and X 3,i are input signals with average power constraints: (x j,i ) P j (57) n for j Î {,, 3}. Z,i and Z 3,i are independent and identically distributed (i.i.d) zero mean Gaussian noise components with unit powers, i.e., Z j,i N (, ) for j Î {, 3}. A. Strong interference regime Here, we extend the results of Theorem 6, i.e., C str and Set, to the Gaussian case. The strong interference conditions of Set, i.e., (34)-(37), for the above Gaussian model become: h 3 (58) P (h 3 ) + ρ P P 3 (h 3 h 3 ) P 3 ( ρ )(h 3 ) (59) P (h 3 ) + ρ P P 3 (h 3 h 3 ) P 3 ( ρ )(h 3 ) (6) m Y X h 3 Z mˆ, mˆ m X h 3 m, m, m3 3 h 3 Z 3 Y X 3 ˆm 3 Figure 4 Gaussian MA-CIFC.

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page of 8 P (h 3 ) + P (h 3 ) + ρ P P 3 (h 3 h 3 ) +ρ P P 3 (h 3 h 3 ) P 3 (h 3 ) (6) where - r u is the correlation coefficient between X u and X 3, i.e., E(X u, X 3 ) ρ u Pu P 3 for u Î {, }. Theorem 8: For the Gaussian MA-CIFC satisfying conditions (58)-(6), the capacity region is given by C G ρ,ρ :ρ +ρ {(R, R, R 3 ):R, R, R 3 R 3 θ(p 3 ( ρ ρ )) (6) R + R 3 θ R + R 3 θ (P + h 3 P 3( ρ )+h 3ρ P P 3 ) (63) (P + h 3 P 3( ρ )+h 3ρ P P 3 ) (64) R + R + R 3 ( θ P + P + h 3 P )} (65) 3 +h 3 P3 (ρ P + ρ P ) where, to simplify notation, we define θ(x). log( + x). (66) Remark 9: Condition (58) implies that Tx3 causes strong interference at Rx. This enables Rx to decode m 3. Moreover, (59)-(6) provide strong interference conditions at Rx3, under which all messages can be decoded in Rx3 in a MAC fashion. Proof: The achievability part follows from C str in Theorem 6 by evaluating (38)-(4) with zero mean jointly Gaussian channel inputs X, X and X 3. That is, X N (, P ), X N (, P ), and X 3 N (, P 3 ), where E(X,X ), E(X, X 3 ) ρ P P 3, and E(X, X 3 ) ρ P P 3. The converse proof is based on reasoning similar to that in [6] and is provided in Appendix C. It is noted that the channel parameters, i.e., P,P,P 3, h 3,h 3,h 3, must satisfy (58)-(6) for all ρ, ρ :ρ + ρ, to numerically evaluate the C G using (6)-(65). Here, we choose P P P 3 6,h 3 h 3 h 3.5 which satisfy strong interference conditions (58)-(6); hence, the regions are derived under strong interference conditions. Figure 5 shows the capacity region for the Gaussian MA-CIFC of Theorem 8, for P P P 3 6,and h 3 h 3 h 3.5, wherer r is fixed in each surface. The r r region corresponds to the no cooperation case, where the channel inputs are independent. It can be seen that as r r increases, the bound.5 ρ ρ ρ ρ.4 R 3.5 ρ ρ.6.5.5 R.5.5.5 R.5 Figure 5 The capacity region for the Gaussian MA-CIFC for fixed r r under the strong interference conditions.

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 on R 3 becomes more restrictive while the sum-rate bounds become looser; because Tx3 dedicates parts of its power for cooperation. This means that, as Tx3 allocates more power to relay m, m by increasing r r, R and R improve, while R 3 degrades due to the less power allocated to transmit m3. The capacity region for this channel is the union of all the regions obtained for different values of r and r, satisfying ρ + ρ. This union is shown in Figure 6. In order to better perceive the effect of cooperation, we let R in Figure 7. It is seen that by increasing r r, the bound on R + R3 becomes looser and R improves, while R3 decreases due to the more power dedicated for cooperation. B. Weak interference regime Now, we consider the Gaussian MA-CIFC with weak interference at the primary receiver (Rx), which means h3. We remark that, since there is no cooperation between receivers, the capacity region for this channel is the same as the one with the same marginal outputs p(yn xn, xn, xn3 ) and p(y3n xn, xn, xn3 ). Hence, we can state the following useful lemma. Lemma : The capacity region of a Gaussian MACIFC, defined by (55) and (56) when h 3, is the same as the capacity region of a Gaussian MA-CIFC with the following channel outputs: Page of 8 Y,i X,i + X,i + h3 Y3,i + Z,i (67) Y3,i h3 X,i + h3 X,i + Y3,i (68) where Y 3,i X 3,i + Z 3,i and Z,i N (, h3 ). Therefore, the degradedness condition in (9) holds for the Gaussian MA-CIFC when h3. Proof: The proof follows from [6, Lemma 3.5]. Next, we use the inner bound of Theorem and the outer bound of Theorem 4 to derive the capacity region, which shows that the capacity-achieving scheme in this case consists of DPC at the cognitive transmitter and treating interference as noise at both receivers. Theorem 9: For the Gaussian MA-CIFC, defined by (55) and (56), when h3, the capacity region is given by CG {(R, R, R3 ) : R, R, R3 ρ,ρ :ρ +ρ (69) R3 θ (P3 ( ρ ρ )) R θ ( P + h3 ρ P3 ) h3 P3 ( ρ ρ ) + (7).5 R3.5.5.5.5.5 R.5.5 Figure 6 The capacity region for the Gaussian MA-CIFC under the strong interference conditions. R

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page of 8.5 ρ ρ ρ ρ increases R 3.5 ρ ρ.5.5.5.5 R Figure 7 The capacity region for the Gaussian MA-CIFC under the strong interference conditions when R. ( ( P + h 3 ρ P3 ) ) R θ h 3 P 3( ρ ρ )+ (7) R + R (( ) ( ) )} P + h 3 ρ P3 + P + h 3 ρ P3 (7) θ h 3 P 3( ρ ρ )+ where θ( ) is defined in (66). Remark : By evaluating () with jointly Gaussian channel inputs, one can easily achieve (69). However, this results in the Gaussian counterparts of the bounds in(7)-().therefore,someconditionsarenecessaryto make these bounds redundant, similar to the ones in (3)-(33). However, we show that () is also evaluated to (69), if we apply DPC with appropriate parameters. Hence, conditions (3)-(33) are unnecessary in the Gaussian case. This means that DPC completely mitigates the effects of interference for the Tx3-Rx3 pair and leaves the link between them interference-free for fixed values of r, r. Consequently, C G is independent of h 3 and h 3. Remark : If we omit Tx, the model reduces to a CIFC and by setting P ρ R,C G coincides with the capacity region of the Gaussian CIFC with weak interference, which was characterized in [6, Lemma 3.6]. Proof: The rate region in Theorem can be extended to the discrete-time Gaussian memoryless case with continuous alphabets by standard arguments [3]. Hence, it is sufficient to evaluate (3)-(6) and () with an appropriate choice of input distribution to reach (69)-(7). Let R,M, andt, sincetxand Tx have no common information. Also, let U and V be deterministic constants. We choose zero mean jointly Gaussian channel inputs X, X and X 3. In fact, X j N (, P j ) for j Î {,,3}, where E(X,X ), E(X, X 3 ) ρ P P 3,andE(X, X 3 ) ρ P P 3.Noting the p.m.f (3), consider the following choice of input distribution for certain { ρ, ρ :ρ + ρ }: X N (, P ), X N (, P ) W X 3 + α X + α X, X 3 N (, ( ρ ρ )P 3) X 3 X 3 + ρ P 3 P 3 X + ρ X P P (73) Therefore, (3)-(6) are easily evaluated to (7)-(7). In Appendix D, we derive (69) by evaluating () with appropriate parameters. The converseprooffollowsby applying the bounds in the proof of Theorem 4 to the Gaussian case and utilizing Entropy Power Inequality (EPI) [3,7]. A detailed converse proof is provided in Appendix D.

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page 3 of 8 Remark : According to Theorems 8 and 9, jointly Gaussian channel inputs X, X and X 3 are optimal for the Gaussian MA-CIFC under the strong and weak interference conditions, determined in the above theorems. Figure 8 shows the capacity region for the Gaussian MA-CIFC of Theorem 9, for P P P 3 6,and h 3.55, where r r is fixed in each surface. It is noted that the capacity region is independent of h 3 and h 3.Ther r region corresponds to the no cooperation case, where channel inputs are independent. We see that when Tx3 dedicates parts of its power for cooperation, i.e., r r.5, the rates of the primary users (R,R ) increase, while R 3 decreases. The capacity region for this channel is the union of all the regions obtained for different values of r and r satisfying, ρ + ρ, which is shown in Figure 9. Similar to Figure 7, we investigate the capacity region for R in Figure in the weak interference regime. It is seen that, when Tx3 dedicates more power for cooperation by increasing r r, R improves and R 3 decreases. 7. Conclusion We investigated a cognitive communication network where a MAC with common information and a pointto-point channel share a same medium and interfere with each other. For this purpose, we introduced Multiple Access-Cognitive Interference Channel (MA- CIFC)bymergingatwo-userMACasaprimarynetwork and a cognitive transmitter-receiver pair in which the cognitive transmitter knows the message being sent by all of the transmitters in a non-causal manner. We analyzed the capacity region of MA-CIFC by deriving the inner and outer bounds on the capacity region of this channel. These bounds were proved to be tight in some special cases. Therefore, we determined the optimal strategy in these cases. Specifically, in the discrete memoryless case, we established the capacity regions for aclassofdegraded MA-CIFC and also under two sets of strong interference conditions. We also derive strong interference conditions for a network with k primary users. Further, we characterized the capacity region of the Gaussian MA-CIFC in the weak and strong interference regimes. We showed that DPC at the cognitive transmitter and treating interference as noise at the receivers, i.e., an oblivious primary receiver, are optimal in the weak interference. However, the receivers have to decode all messages when the interference is strong enough, which requires an aware primary receiver. Appendix A Proofs of Theorems, and 3 Outline of the proof for Theorem : We propose the following random coding scheme, which contains superposition coding and the technique of [6] in defining.5 ρ ρ R 3.5..4.6 R.8..4 ρ ρ.5.4..8.6 R.4. Figure 8 The capacity region for the weak Gaussian MA-CIFC for fixed r r.

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page 4 of 8.5 R 3.5.5 R.5.5.5.5 R.5 Figure 9 The capacity region for the weak Gaussian MA-CIFC..5 ρ ρ ρ ρ increases R 3.5 ρ ρ.5..4.6.8..4.6.8 R Figure The capacity region for the weak Gaussian MA-CIFC when R.

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page 5 of 8 auxiliary RVs, i.e., U and V. The cognitive receiver (Rx3) decodes the interfering signals caused by the primary messages (m, m ), while the primary receiver (Rx) does not decode the interference from the cognitive user s message (m 3 ) and treats it as noise. Codebook Generation: Fixajointp.m.fas().Generate nr i.i.d t n sequences, according to the probability n p(t i ). Index them as t n (m )wherem [, nr ]. For each t n (m ), generate nr i.i.d (u n, x n ) sequences, each n with probability p(u i, x,i t i ). Index them as (u n (m, m ), x n (m, m )) where m [, nr ]. Similarly, for each t n (m ), generate nr i.i.d (v n, x n n ) sequences, each with probability p(v i, x,i t i ). Index them as (v n (m, m ), x n (m, m )) where m [, nr ]. For each (t n (m ), u n (m, m ), x n (m, m ), v n (m, m ), x n (m, m )), generate nr 3 i.i.d x n 3 sequences, according to n p(x 3,i t i, u i, x,i, v i, x,i ). Index them as x n 3 (m, m, m, m 3 ) where m 3 [, nr 3]. Encoding: In order to transmit the messages (m, m, m, m 3 ), Txj sends x n j (m, m j ) for j Î {,} and Tx3 sends x n 3 (m, m, m, m 3 ). Decoding: Rx: After receiving y n,rxlooksforauniquetriple ( ˆm, ˆm, ˆm ) and some ˆm such that (y n, tn ( m ), u n ( m, m ), x n ( m, m ), v n ( m, m ), x n ( m, m )) A n (Y, T, U, X, V, X ). For large enough n with arbitrarily high probability, ( ˆm, ˆm, ˆm ) (m, m, m ) if (3)-(6) hold. Rx3: After receiving y n 3, Rx3 finds a unique index ˆm 3 and some triple ( ˆm, ˆm, ˆm ) such that ( y n 3, x n 3 ( m, m, m, m 3 ), t n ( m ), u n ( m, m ), x n ( m, m ), v n ( m, m ), x n ( m, m ) ) A n (Y 3, X 3, T, U, X, V, X ). With arbitrary high probability, ˆm 3 m 3 if n is large enough and (), (7)-() hold. This completes the proof. Outline of the proof for Theorem : Our proposed random coding scheme, in the encoding part, contains the methods of Theorem and GP binning at the cognitive transmitter (Tx3) which is used at Tx3 to cancel the interference caused by m, m, m at Rx3. In the decoding part, both receivers decode only their intended messages, treating the interference as noise. Therefore, unlike the decoding part of Theorem, Rx3 decodes only its own message (m3), while treating the other signals as noise. Codebook Generation: Fixajointp.m.fas(3).Generate t n (m ), u n (m, m ), x n (m, m ), v n (m, m ), x n (m, m ) codewords based on the same lines as in the codebook generation part of Theorem. Then, generate n(r 3+L) i.i. d w n sequences. Index them as w n (m 3,l) where m 3 [, nr 3] and l Î [, nl ]. Encoding: In order to transmit the messages (m,m, m,m 3 ), Txj sends x n j (m, m j ) for j Î {,}. Tx3 (the cognitive transmitter), in addition to m 3, knows m, m and m. Hence, knowing codewords t n, u n, x n, vn, x n,tosend m 3, it seeks an index l such that (w n (m 3, l), t n (m ), u n (m, m ), x n (m, m ), v n (m, m ), x n (m, m )) A n (W, T, U, X, V, X ). Ifthereismorethanonesuchindex,Tx3picksthe smallest. If there is no such codeword, it declares an error. Using covering lemma [7], it can be shown that there exists such an index l with high enough probability if n is sufficiently large and L I(W; T, U, X, V, X ). (74) Then, Tx3 sends x n 3 generated according to n p(x 3,i t i, u i, x,i, v i, x,i ). Decoding: The decoding procedure at Rx is similar to Theorem and the error in this receiver can be bounded if (3)-(6) hold. Rx3: After receiving y n 3, Rx3 finds a unique index ˆm 3 for some index ˆl such that ( ( m )) y3 n, wn 3, l A n (Y 3, W). For large enough n, the probability of error can be made sufficiently small if R 3 + L I(W; Y 3 ). (75) Combining (74) and (75) results in (). This completes the proof. Outline of the proof for Theorem 3: We propose the following random coding scheme, which contains superposition coding in the encoding part and simultaneous joint decoding in the decoding part. All messages are common to both receivers, i.e., both receivers decode (m, m, m, m 3 ). Codebook Generation: Fixajointp.m.fas(8).Generate nr i.i.d t n sequences, according to the probability n p(t i ). Index them as t n (m )wherem [, nr ]. Forj

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page 6 of 8 Î {,} and each t n (m ), generate nr j i.i.d x n j sequences, n each with probability p(x j,i t i ). Index them as x n j (m, m j ) where m j [, nr j ]. For each (t n (m ), x n (m, m ), x n (m, m )), generate nr 3 i.i.d x n 3 n sequences, each with probability p(x 3,i t i, x,i, x,i ). Index them as x n 3 (m, m, m, m 3 ) where m 3 [, nr 3]. Encoding: In order to transmit message (m, m, m, m 3 ), Txj sends x n j (m, m j ) for j Î {,} and Tx3 sends x n 3 (m, m, m, m 3 ). Decoding: Rx: After receiving y n,rxlooksforauniquetriple ( ˆm, ˆm, ˆm ) and some ˆm 3 such that (y n, tn ( m ), x n ( m, m ), x n ( m, m ), x n 3 ( m, m, m, m 3 )) A n (Y, T, X, X, X 3 ). For large enough n, with arbitrarily high probability ( ˆm, ˆm, ˆm ) (m, m, m ) if R + R 3 I(X, X 3 ; Y X, T) (76) R + R 3 I(X, X 3 ; Y X, T) (77) R + R + R + R 3 I(X, X, X 3 ; Y ). (78) Rx3: Similarly, after receiving y n 3, Rx3 finds a unique index ˆm 3 and some triple ( ˆm, ˆm, ˆm ) such that (y n 3, tn ( ˆm ), x n ( ˆm, ˆm ), x n ( ˆm, ˆm ), x n 3 ( ˆm, ˆm, ˆm, ˆm 3 )) A n ε (Y 3, T, X, X, X 3 ). With the arbitrary high probability ˆm 3 m 3,ifn is large enough and R 3 I(X 3 ; Y 3 X, X, T) (79) R + R 3 I(X, X 3 ; Y 3 X, T) (8) R + R 3 I(X, X 3 ; Y 3 X, T) (8) R + R + R + R 3 I(X, X ; Y 3 X ). (8) This completes the proof. Appendix B Proof of Theorem 7 Since achievability directly follows from Theorem 3, we proceed to the converse part. We assume a code with the properties indicated in the converse proof of Theorem 6 and define T n as (4). Four bounds in C str are the same as the bounds in C str, which are shown in the converse proof of Theorem 6. Therefore, it is necessary to prove the three bounds. Moreover, similar to Lemma, we have: Lemma 3: If (48)-(5) hold for all distributions that factor as (8), then I(X n ; Yn Xn, Tn ) I(X n ; Yn 3 Xn, Tn ) (83) I(X n ; Yn Xn, Tn ) I(X n ; Yn 3 Xn, Tn ) (84) I(X n ; Xn ; Yn ) I(Xn ; Yn Xn 3 ). (85) We next prove the remaining three bounds of Theorem 7. Applying Fano s inequality, similar to (45), we have: n(r + R 3 ) n(δ n + δ 3n ) I(M, X n ; Yn M, M, X n )+I(Xn 3 ; Yn 3 M, M, M, X n, Xn ) I(X n ; Yn Tn, M, X n )+I(Xn 3 ; Yn 3 Tn, M, M, X n, Xn ) I(X n ; Yn 3 Tn, M, X n )+I(Xn 3 ; Yn 3 Tn, M, M, X n, Xn ) (c) I(M, X n, Xn 3 ; Yn 3 Tn, M, X n ) (d) I(X,i, X 3,i ; Y 3,i X,i, T i ) (86) where follows from (4) and the deterministic relation between X n and M, is obtained from (83), (c) is due to the fact that mutual information is non-negative, and (d) follows from the facts that conditioning does not increase entropy and the channel is memoryless. Using a similar approach and condition (84), it can be shown that n(r + R 3 ) n(δ n + δ 3n ) I(X,i, X 3,i Y 3,i X,i, T i ). (87) Finally, using (47) we obtain the last sum-rate bound as: n(r + R + R + R 3 ) n(δ n + δ n + δ n + δ 3n ) I(M, M, M, X n, Xn ; Yn ) +I(X3 n ; Yn 3 M, M, M, X n, Xn ) I(X n, Xn ; Yn )+I(Xn 3 ; Yn 3 Xn, Xn ) I(X n, Xn ; Yn 3 )+I(Xn 3 ; Yn 3 Xn, Xn ) I(X n, Xn, Xn 3 ; Yn 3 ) I(X,i,X,i X 3,i ; Y 3,i ) (88) where is obtained from the deterministic relation between Xj n and (M, M j )forjî {,}, and follows from (85). Using a standard time-sharing argument for (44)-(47) and (86)-(88) completes the proof.

Mirmohseni et al. EURASIP Journal on Wireless Communications and Networking, :5 http://jwcn.eurasipjournals.com/content///5 Page 7 of 8 Appendix C Proof of the converse part for Theorem 8 Foranyratetriple(R, R R 3 ) C, Rx is able to decode m and m reliably. Then, Rx is able to construct Ỹ 3 Y X X h 3 + h 3 X + h 3 X h 3 X + h 3 X + X 3 + h 3 Z. If condition (58) holds, then Ỹ 3 is a less noisy version of Y 3. Since Rx3 has to decode m 3, Rx can decode m 3 via Ỹ 3 with no rate penalty. Therefore, (R, R, R 3 )iscontained in the capacity region of a three-user MAC with common information [9] at Rx, where R and R are the common rates between Tx-Tx3 and Tx-Tx3, respectively; R 3 is the private rate for Tx3; and the private rates for Tx and Tx are zero. From the maximum entropy theorem [3], this region is largest for Gaussian inputs and is evaluated to (63)-(65). The bound in (6) follows by applying the standard methods as in (44). Appendix D Detailed proof for Theorem 9 First, we derive (69) by evaluating () with appropriate parameters. Consider the mapping in (73). We choose ) P 3 P 3 α β (h 3 + ρ ρ (89) P P P 3 P 3 α β(h 3 + ρ ) ρ (9) P P β P 3 ( ρ ρ ) P 3 ( ρ ρ )+ (9) It is noted that (9) makes ( - b)x 3 - bz 3 and X 3 + Z 3 uncorrelated and hence independent since they are jointly Gaussian. Using () and (73), we obtain I(W; Y 3 ) I(W; T, U, X, V, X )h(w X, X ) h(w Y 3 ) (9) h(x 3) h(w Y 3 ). Now, we evaluate h(w Y 3 ): h(w Y 3 )h(x P 3 P 3 3 +(α + ρ )X +(α + ρ )X Y 3 ) P P h(x P 3 P 3 3 +(α + ρ )X +(α + ρ )X βy 3 Y 3 ) P P h(( β)x p3 3 βz 3 (h 3 + ρ )X P P 3 +(h 3 + ρ )X + X 3 + Z 3 P h(( β)x 3 βz 3 X 3 + Z 3 ) h(x 3 X 3 + Z 3 ) (93) where follows from (89), (9), and P3 P3 Y 3 + ρ )X +(h 3 + ρ )X + X 3 P P + Z 3 is obtained because, by applying (9), ( β)x 3 βz 3, X 3 + Z 3, X, X and X are jointly independent. By applying (93) to (9), we can evaluate () as: h(x 3) h(w Y 3 )h(x 3) h(x 3 X 3 + Z 3 ) I(X 3; X 3 + Z 3 )θ(p 3 ( ρ ρ )) This completes the achievability proof. Converse: Utilizing the power constraint (57), we derive the bounds of R o for the Gaussian MA-CIFC when h 3 <. It can be easily shown that we can set T, since R and M. Applying the degradedness condition (9) to (9), which holds due to Lemma, we obtain: R 3 I(X 3 ; Y 3, Y U, X, V, X )I(X 3 ; Y 3 U, X, V, X ) h(y 3 U, X, V, X ) log πe. (94) Using the fact that conditioning does not increase the entropy and (57) with j 3 yields: h(y 3 U, X, V, X ) h(y 3 U, X, V, X, X 3 ) log πe h(y 3 U, X, V, X )h(x 3 + Z 3 U, X, V, X, X 3 ) h(x 3 + Z 3 ) log πe(p 3 +). Hence, there exists l such that h(y 3 U, X, V, X ) log πe(λp 3 +). (95) Combining (94) and (95) results in R 3 I(X 3 ; Y 3 U, X, V, X ) log(λp 3 +) (96) for some l. Now, considering that the Gaussian distribution maximizes the entropy of a RV for a given value of the second moment, we have: h(y 3 U, X, V, X ) h(x 3 + Z 3 X, X ) log πe(p 3( ρ ρ )+)(97) where ρ, ρ :ρ + ρ, ρ E(X, X 3 ) P P 3, and ρ E(X, X 3 ). Considering (95) and (97), we P P 3 obtain: λ ρ ρ Moreover, since (69) is achievable and R 3 does not appear in (7)-(7), it can be shown that