- Research
- Open Access
- Published:

# Capacity bounds for multiple access-cognitive interference channel

*EURASIP Journal on Wireless Communications and Networking*
**volume 2011**, Article number: 152 (2011)

## Abstract

Motivated by the uplink scenario in cellular cognitive radio, this study considers a communication network in which a point-to-point channel with a cognitive transmitter and a Multiple Access Channel (MAC) with common information share the same medium and interfere with each other. A Multiple Access-Cognitive Interference Channel (MA-CIFC) is proposed with three transmitters and two receivers, and its capacity region in different interference regimes is investigated. First, the inner bounds on the capacity region for the general discrete memoryless case are derived. Next, an outer bound on the capacity region for full parameter regime is provided. Using the derived inner and outer bounds, the capacity region for a class of *degraded* MA-CIFC is characterized. Two sets of strong interference conditions are also derived under which the capacity regions are established. Then, an investigation of the Gaussian case is presented, and the capacity regions are derived in the *weak* and *strong* interference regimes. Some numerical examples are also provided.

## 1. Introduction

Interference avoidance techniques have traditionally been used in wireless networks wherein multiple source-destination pairs share the same medium. However, the broadcasting nature of wireless networks may enable cooperation among entities, which ensures higher rates with more reliable communication. On the other hand, due to the increasing number of wireless systems, spectrum resources have become scarce and expensive. The exponentially growing demand for wireless services along with the rapid advancements in wireless technology has lead to cognitive radio technology which aims to overcome the spectrum inefficiency problem by developing communication systems that have the capability to sense the environment and adapt to it [1].

In overlay cognitive networks, the cognitive user can transmit simultaneously with the non-cognitive users and compensate for the interference by cooperation in sending, i.e., relaying, the non-cognitive users' messages [1]. From an information theoretic point of view, Cognitive Interference Channel (CIFC) was first introduced in [2] to model an overlay cognitive radio and refers to a two-user Interference Channel (IFC) in which the cognitive user (secondary user) has the ability to obtain the message being transmitted by the other user (primary user), either in a non-causal or in a causal manner. An achievable rate region for the non-causal CIFC was derived in [2], by combining the Gel'fand-Pinsker (GP) binning [3] with a well-known simultaneous superposition coding scheme (rate splitting) applied to IFC [4]. For the non-causal CIFC, where the cognitive user has non-causal full or partial knowledge of the primary user's transmitted message several achievable rate regions and capacity results in some special cases have been established [5–14]. More recently a *three-user* cognitive radio network with one primary user and two cognitive users is studied in [15, 16], where an achievable rate region is derived for this setup based on rate splitting and GP binning.

In the interference avoidance-based systems, i.e., when the communication medium is interference-free, uplink transmission is modeled with a Multiple Access Channel (MAC) whose capacity region has been fully characterized for independent transmitters [17, 18] as well as for the transmitters with common information [19]. Recently, taking the effects of interference into account in the uplink scenario, a MAC and an IFC have been merged into one setup by adding one more transmit-receive pair to the communication medium of a two-user MAC [20, 21], where the channel inputs at the transmitters are independent and there is no cognition or cooperation.

In this paper, we introduce Multiple Access-Cognitive Interference Channel (MA-CIFC) by providing the transmitter of the point-to-point channel with *cognition capabilities* in the uplink with interference model. Moreover, transmitters of MAC have common information that enables cooperation among them. As shown in Figure 1, the proposed channe consists of three transmitters and two receivers: two-user MAC with common information as the primary network and a point-to-point channel with a cognitive transmitter that knows the message being sent by all of the transmitters in a non-causa manner. A physical example of this channel is the coexistence of cognitive users with the licensed primary users in a cellular or satellite uplink transmission, where the cognitive radios by their abilities exploit side information about the environment to maintain or improve the communication of primary users while also achieving some spectrum resources for their own communication. In this scenario, the primary non-cognitive users can be *oblivious* to the or *aware* of the cognitive users [1]. When the non-cognitive user is oblivious to the cognitive user's presence, its receiver's decoding process is independent of the interference caused by the cognitive user's transmission. In fact, the primary receiver treats interference as noise. However, in the aware cognitive user's scenario, the decoding process at the primary receiver can be adapted to improve its own rate. For example, the primary receiver can decode the cognitive user's message and cancel the interference when the interfering signal is strong enough. If the multi-antenna capability is available at the primary receiver, it can also reduce or increase the interfering signal by beam-steering, which results in the occurrence of the weak or strong interference regimes [1].

To analyze the capacity region of MA-CIFC, we first derive three inner bounds on the capacity region (achievable rate regions). The first two bounds assume an oblivious primary receiver, which does not decode the cognitive user's message but treats it as noise. Two different coding schemes are proposed based on the superposition coding, the GP binning and the method of [6] in defining auxiliary Random Variables (RVs). Later, we show that these strategies are optimal for a *degraded* MA-CIFC and also in the Gaussian *weak* interference regime. In the third achievability scheme, we consider an aware primary receiver and obtain an inner bound on the capacity region based on using superposition coding in the encoding part and allowing both receivers to decode all messages with simultaneous joint decoding in the decoding part. This strategy is capacity-achieving in the strong interference regime. Next, we provide a general outer bound on the capacity region and derive conditions under which the first achievability scheme achieves capacity for the *degraded* MA-CIFC. We continue the capacity results by the derivation of two sets of strong interference conditions, under which the third inner bound achieves capacity. Further, we compare these two sets of conditions and identify the weaker set. We also extend the strong interference results to a network with *k* primary users.

Moreover, we consider the Gaussian case and find capacity results for the Gaussian MA-CIFC in both the *weak* and *strong* interference regimes. We use the second derived inner bound to show that the capacity-achieving scheme in weak interference consists of Dirty Paper Coding (DPC) [22] at the cognitive transmitter and treating interference as noise at both receivers. We also provide some numerical examples.

The rest of the paper is organized as follows. Section 2 introduces MA-CIFC model and the notations. Three inner bounds and an outer bound on the capacity region are derived in Section 3 and Section 4, respectively, for the discrete memoryless MA-CIFC. Section 5 presents the capacity results for the discrete memoryless MA-CIFC in three special cases. In Section 6, the Gaussian MA-CIFC is investigated. Finally, Section 7 concludes the paper.

## 2. Channel models and preliminaries

Throughout the paper, upper case letters (e.g. *X*) are used to denote RVs and lower case letters (e.g. *x*) show their realizations. The probability mass function (p.m.f) of a RV *X* with alphabet set \mathcal{X} is denoted by *p*_{
X
}(*x*), where subscript *X* is occasionally omitted. {A}_{\epsilon}^{n}\left(X,Y\right) specifies the set of ϵ-strongly, jointly typical sequences of length *n*. The notation {X}_{i}^{j} indicates a sequence of RVs (*X*_{
i
}, *X*_{i+1}, ..., *X*_{
j
}), where *X*^{j}is used instead of {X}_{1}^{j}, for brevity. \mathcal{N}\left(0,{\sigma}^{2}\right) denotes a zero mean normal distribution with variance *σ*^{2}.

Consider the MA-CIFC in Figure 2, which is denoted by \left({\mathcal{X}}_{1}\times {\mathcal{X}}_{2}\times {\mathcal{X}}_{3},p\left({y}_{1}^{n},{y}_{3}^{n}|{x}_{1}^{n},{x}_{2}^{n},{x}_{3}^{n}\right),{\mathcal{Y}}_{1}\times {\mathcal{Y}}_{3}\right), where {X}_{1}\in {\mathcal{X}}_{1},{X}_{2}\in {\mathcal{X}}_{2} and {X}_{3}\in {\mathcal{X}}_{3} are channel inputs at Transmitter 1 (Tx1), Transmitter 2 (Tx2) and Transmitter 3 (Tx3), respectively; {Y}_{1}\in {\mathcal{Y}}_{1} and {Y}_{3}\in {\mathcal{Y}}_{3} are channel outputs at Receiver 1 (Rx1) and Receiver 3 (Rx3), respectively; and p\left({y}_{1}^{n},{y}_{3}^{n}|{x}_{1}^{n},{x}_{2}^{n},{x}_{3}^{n}\right) is the channel transition probability distribution. In *n* channel uses, each Tx*j* desires to send a message pair (*m*_{0}, *m*_{
j
}) to Rx1 where *j* ∈ {1,2}, and Tx3 desires to send a message *m*_{3} to Rx3.

*Definition 1:* A \left({2}^{n{R}_{0}},{2}^{n{R}_{1}},{2}^{n{R}_{2}},{2}^{n{R}_{3}},n\right) code for MA-CIFC consists of (i) four independent message sets {\mathcal{M}}_{j}=\left\{1,...,{2}^{n{R}_{j}}\right\}, where *j* ∈ {0, 1, 2, 3}; (ii) two encoding functions at the primary transmitters, {f}_{1}:{\mathcal{M}}_{0}\times {\mathcal{M}}_{1}\mapsto {\mathcal{X}}_{1}^{n} at Tx1 and {f}_{2}:{\mathcal{M}}_{0}\times {\mathcal{M}}_{2}\mapsto {\mathcal{X}}_{2}^{n} at Tx2; (iii) an encoding function at the cognitive transmitter, {f}_{3}:{\mathcal{M}}_{0}\times {\mathcal{M}}_{1}\times {\mathcal{M}}_{2}\times {\mathcal{M}}_{3}\mapsto {\mathcal{X}}_{3}^{n}; and (iv) two decoding functions, {g}_{1}:{\mathcal{Y}}_{1}^{n}\mapsto {\mathcal{M}}_{0}\times {\mathcal{M}}_{1}\times {\mathcal{M}}_{2} at Rx1 and {g}_{3}:{\mathcal{Y}}_{3}^{n}\mapsto {\mathcal{M}}_{3} at Rx3. We assume that the channel is memoryless. Thus, the channel transition probability distribution is given by

The probability of error for this code is defined as

*Definition 2:* A rate quadruple (*R*_{0}, *R*_{1}, *R*_{2}, *R*_{3}) is achievable if there exists a sequence of \left({2}^{n{R}_{0}},{2}^{n{R}_{1}},{2}^{n{R}_{2}},{2}^{n{R}_{s}},n\right) codes with *P*_{
e
}→ 0 as *n* → ∞. The capacity region \mathcal{C}, is the closure of the set of all achievable rates.

## 3. Inner bounds on the capacity region of discrete memoryless MA-CIFC

Now, we derive three achievable rate regions for the general setup. Theorems 1 and 2 assume an oblivious primary receiver (Rx1), which does not decode the cognitive user's message (*m*_{3}) and treats it as noise. The decoding procedure at the cognitive receiver (Rx3) differs in these schemes. In Theorem 1, the cognitive receiver (Rx3) decodes the primary messages (*m*_{0}, *m*_{1}, *m*_{2}), and all the transmitters use superposition coding. However, in Theorem 2, the cognitive receiver (Rx3) also treats the interference from the primary messages (*m*_{0}, *m*_{1}, *m*_{2}) as noise, while the cognitive transmitter (Tx3) uses GP binning to precode its message for interference cancelation at Rx3. We also utilize the method of [6] in defining auxiliary RVs, which helps us to achieve the outer bound in special cases. In fact, we achieve the outer bound of Theorem 4 using the region of Theorem 1 for a class of degraded MA-CIFC in Section 5. The region of Theorem 2 is used in Section 6 to derive the capacity region in the *weak* interference regime. In the scheme of Theorem 3, we consider an aware primary receiver (Rx1) which decodes the cognitive user's message (*m*_{3}). The cognitive receiver (Rx3) also decodes the primary messages (*m*_{0}, *m*_{1}, *m*_{2}). Therefore, this region is obtained based on using superposition coding in the encoding part and by allowing both receivers to decode all messages with simultaneous joint decoding in the decoding part. In Section 5, we show that this strategy is capacity-achieving in the strong interference regime. Proofs are provided in "Appendix A".

*Theorem 1:* The union of rate regions given by

is achievable for MA-CIFC, where the union is over all p.m.fs that factor as

*Theorem 2:* The union of rate regions given by (3)-(6) and

is achievable for MA-CIFC, where the union is over all p.m.fs that factor as

*Theorem 3:* The union of rate regions given by

is achievable for MA-CIFC, where the union is over all p.m.fs that factor as

*Remark 1:* We utilize the region of Theorem 1 in Section 5 to achieve capacity results for a class of degraded MA-CIFC, and the region of Theorem 2 to derive the results for the Gaussian case in Section 6. The region of Theorem 3 is also used to characterize the capacity region under strong interference conditions in Section 5.

## 4. An outer bound on the capacity region of discrete memoryless MA-CIFC

Here, we derive a general outer bound on the capacity region of MA-CIFC which is used to obtain the capacity region for a class of the degraded MA-CIFC in Section 5 and also to find capacity results for the Gaussian MA-CIFC in the weak interference regime in Section 6. Let {\mathcal{R}}_{o}^{1} denote the union of all rate quadruples (*R*_{0}, *R*_{1}, *R*_{2}, *R*_{3}) satisfying (3)-(6) and

where the union is over all p.m.fs that factor as (11).

*Theorem 4:* The capacity region of MA-CIFC satisfies

*Proof:* Consider a \left({2}^{n{R}_{0}},{2}^{n{R}_{1}},{2}^{n{R}_{2}},{2}^{n{R}_{3}},n\right) code with the average error probability of {P}_{e}^{n}\to 0. Define the following RVs for *i* = 1, ..., *n*:

Considering the encoding functions *f*_{1} and *f*_{2}, defined in Definition 1, and the above definitions for auxiliary RVs, we remark that (*X*_{1,i}, *U*_{
i
}) → *T*_{
i
}→ (*X*_{2,i}, *V*_{
i
}) forms a Markov chain. Thus, these choices of auxiliary RVs satisfy the p.m.f (11) of Theorem 4. Now using Fano's inequality [23], we derive the bounds in Theorem 4. For the first bound, we have:

where (a) follows since messages are independent and (b) holds due to Fano's inequality and the fact that conditioning does not increase entropy. Hence,

where (a) is due to the encoding functions *f*_{1}, *f*_{2} and *f*_{3}, defined in Definition 1, and the non-negativity of mutual information, (b) is obtained from the chain rule, (c) follows from the memoryless property of the channel and the fact that conditioning does not increase entropy, and (d) is obtained from (20)-(22).

Now, applying Fano's inequality and the independence of the messages, we can bound *R*_{1} as:

where (a) follows from the chain rule and the encoding functions *f*_{1} and *f*_{2}, and (b) from (20)-(22). Similarly, we can show that

Next, based on similar arguments, we bound *R*_{1} + *R*_{2} as

The last sum-rate bound can be derived as follows:

where (a) follows since conditioning does not increase entropy. Using the standard time-sharing argument for (24)-(28) completes the proof.

## 5. Capacity results for discrete memoryless MA-CIFC

In this section, we characterize the capacity region of MA-CIFC under specific conditions. First, we consider a class of degraded MA-CIFC and derive conditions under which the inner bound in Theorem 1 achieves the outer bound of Theorem 4. Next, we investigate the strong interference regime by deriving two sets of strong interference conditions under which the region of Theorem 3 achieves capacity. We also compare these two sets of conditions and identify the weaker set. Finally, we extend the strong interference results to a network with *k* primary users.

### A. Degraded MA-CIFC

Now, we characterize the capacity region for a class of MA-CIFC with a *degraded* primary receiver. We define MA-CIFC with a degraded primary receiver as a MA-CIFC where *Y*_{1} and *X*_{3} are independent given *Y*_{3}, *X*_{1}, *X*_{2}. More precisely, the following Markov chain holds:

or equivalently, *X*_{3} → (*X*_{1}, *X*_{2}, *Y*_{3}) → *Y*_{1} forms a Markov chain. This means that the primary receiver (Rx1) observes a degraded or noisier version of the cognitive user's signal (Tx3) compared with the cognitive receiver (Rx3).

Assume that the following conditions are satisfied for MA-CIFC over all p.m.fs that factor as (11):

Under these conditions, the cognitive receiver (Rx3) can decode the messages of the primary users with no rate penalty. If MA-CIFC with a degraded primary receiver satisfies conditions (30)-(33), the region of Theorem 1 coincides with {\mathcal{R}}_{o}^{1} and achieves capacity, as stated in the following theorem.

*Theorem 5:* The capacity region of MA-CIFC with a degraded primary receiver, defined in (29), satisfying (30)-(33) is given by the union of rate regions satisfying (2)-(6) over all joint p.m.fs (11).

*Remark 2:* The messages of the primary users (*m*_{0}, *m*_{1}, *m*_{2}) can be decoded at Rx3 under conditions (30)-(33). Therefore, Rx3-Tx3 achieves the rate in (2). Moreover, we can see that due to the degradedness condition in (29), treating interference as noise at the primary receiver (Rx1) achieves capacity. We show in Section 6 that, in the Gaussian case the capacity is achieved by using the region of Theorem 2 based on DPC (or GP binning), where the cognitive receiver (Rx3) does not decode the primary messages and conditions (30)-(33) are not necessary.

*Proof:* __Achievability:__ The proof follows from the region of Theorem 1. Using the condition in (30), the sum of the bounds in (2) and (3) makes the bound in (7) redundant. Similarly, conditions (31)-(33), along with the bound in (2), make the bounds in (8)-(10) redundant and the region reduces to (2)-(6).

__Converse:__ To prove the converse part, we evaluate {\mathcal{R}}_{o}^{1} of Theorem 4 with the degradedness condition in (29). It is noted that the p.m.f of Theorem 5 is the same as the one for {\mathcal{R}}_{o}^{1}. Moreover, the bounds in (3)-(6) are equal for both regions. Hence, it is only necessary to show the bound in (2). Considering (19), we obtain:

where (a) is obtained by applying the degradedness condition in (29). This completes the proof.

### B. Strong interference regime

Now, we derive two sets of strong interference conditions under which the region of Theorem 3 achieves capacity. First, assume that the following set of strong interference conditions, referred to as *Set1*, holds for all p.m.fs that factor as (18):

In fact, under these conditions, interfering signals at the receivers are strong enough that all messages can be decoded by both receivers. Condition (34) implies that the cognitive user's message (*m*_{3}) can be decoded at Rx1, while conditions (35)-(37) guarantee the decoding of the primary messages (*m*_{0}, *m*_{1}, *m*_{2}) along with *m*_{3} at Rx3 in a MAC fashion.

*Theorem 6:* The capacity region of MA-CIFC satisfying (34)-(37) is given by:

*Remark 3:* The message of the cognitive user (*m*_{3}) can be decoded at Rx1, under condition (34) and (*m*_{0}, *m*_{1}, *m*_{2}) can be decoded at Rx3 under conditions (35)-(37). Hence, the bound in (38) gives the capacity of a point-to-point channel with message *m*_{3} with side-information *X*_{1}, *X*_{2} at the receiver. Moreover, (38)-(41) with condition (34) give the capacity region for a three-user MAC with common information where *R*_{1} and *R*_{2} are the common rates, *R*_{3} is the private rate for Tx3, and the private rates for Tx1 and Tx2 are zero.

*Remark 4:* If we omit Tx2, i.e., {X}_{2}=\varnothing, and Tx2 has no message to transmit, i.e., *R*_{2} = 0, the model reduces to a CIFC, and {\mathcal{C}}_{1}^{\mathsf{\text{str}}} coincides with the capacity region of the strong interference channel with unidirectional cooperation (or CIFC), which was characterized in [8, Theorem 5]. It is noted that in this case, the common message can be ignored, i.e., T=\varnothing and *R*_{0} = 0.

*Proof:* __Achievability:__ Considering (35)-(37), the proof follows from Theorem 3.

__Converse:__ Consider a \left({2}^{n{R}_{0}},{2}^{n{R}_{1}},{2}^{n{R}_{2}},{2}^{n{R}_{3}},n\right) code with an average error probability of {P}_{e}^{n}\to 0. Define the following RV for *i* = 1, ..., *n*:

It is noted that due to the encoding functions *f*_{1}, *f*_{2} and *f*_{3}, defined in Definition 1, the independence of messages, and the above definitions for *T*^{n}, RVs satisfy the p.m.f (18) of Theorem 6. First, we provide a useful lemma which we need in the proof of the converse part.

*Lemma 1:* If (34) holds for all distributions that factor as (18), then

*Proof:* The proof relies on the results in [24, Proposition 1] and [25, Lemma]. By redefining *X*_{2} = *X*_{3}, *Y*_{2} = *Y*_{3}, *X*_{1} = (*X*_{1}, *X*_{2}, T) in [8, Lemma 5], the proof follows.

Now, using Fano's inequality [23], we derive the bounds in Theorem 6. Using (23) provides:

where (a) is due to (42) and the encoding functions *f*_{1}, *f*_{2} and *f*_{3}, defined in Definition 1, (b) follows from two facts; conditioning does not increase entropy and \left({M}_{1},{M}_{2},{M}_{3}\right)\to \left({X}_{1}^{n},{X}_{2}^{n},{X}_{3}^{n}\right)\to {Y}_{3}^{n} forms a Markov chain, (c) is obtained from the chain rule, and (d) follows from the memoryless property of the channel and the fact that conditioning does not increase entropy.

Now, applying Fano's inequality and the independence of the messages, we can bound *R*_{1} + *R*_{3} as

where (a) follows from encoding functions *f*_{1}, *f*_{2} and *f*_{3}, (b) follows from (42) and the fact that {M}_{3}\to \left({X}_{1}^{n},{X}_{2}^{n},{X}_{3}^{n}\right)\to {Y}_{3}^{n} forms a Markov chain, (c) is obtained from (43), (d) follows from the chain rule, and (e) follows from the memoryless property of the channel and the fact that conditioning does not increase entropy.

Applying similar steps, we can show that,

Finally, the sum-rate bound can be obtained as

where (a) follows from steps (a)-(c) in (45), (b) is due to the fact that \left({M}_{1},{M}_{2}\right)\to \left({X}_{1}^{n},{X}_{2}^{n},{X}_{3}^{n}\right)\to {Y}_{1}^{n} forms a Markov chain, and (c) follows from the memoryless property of the channel and the fact that conditioning does not increase entropy. Using a standard time-sharing argument for (44)-(47) completes the proof.

Next, we derive the second set of strong interference conditions, called *Set2*, under which the region of Theorem 3 is the capacity region. For all p.m.fs that factor as (18), *Set2* includes (34) and the following conditions:

*Remark 5:* Similar to the condition *Set1*, under these conditions interfering signals at the receivers are strong enough that all messages can be decoded by both receivers. The first condition in (34) is equal in the two sets under which the cognitive user's message (*m*_{3}) can be decoded at Rx1. However, conditions (48)-(50) imply that the primary messages (*m*_{0}, *m*_{1}, *m*_{2}) can be decoded at Rx3 in a MAC fashion, while in *Set1*, they can be decoded along with *m*_{3}.

*Theorem 7:* The capacity region of MA-CIFC, satisfying (34) and (48)-(50), referred to as {\mathcal{C}}_{2}^{\mathsf{\text{str}}}, is given by the union of rate regions satisfying (14)-(17) over all p.m.fs that factor as (18).

*Proof:* See "Appendix B".

*Remark 6:* Similar to Remark 4, by omitting \mathsf{\text{Tx}}2\phantom{\rule{2.77695pt}{0ex}}\left(T={X}_{2}=\varnothing ,\phantom{\rule{2.77695pt}{0ex}}{R}_{0}={R}_{2}=0\right), the model reduces to a CIFC. Moreover, {\mathcal{C}}_{2} and *Set2* reduce to the capacity region and strong interference conditions which have been derived in [13] for non-causal CIFC.

*Remark 7 (Comparison of two sets of conditions):* In the strong interference conditions of *Set1*, the first condition in (34) is used in the converse part, while (35)-(37) are used to reduce the inner bound to {\mathcal{C}}_{1}^{\mathsf{\text{str}}}. However, all the conditions of *Set2* are utilized to prove the converse part. Now, we compare the conditions in these two sets. We can write (35) as

Considering (34), it can be seen that *I*_{diff} ≥ 0. Hence, condition (35) implies condition (48), but not vice versa. Similar conclusions can be drawn for other conditions of these two sets. Therefore, *Set1* implies *Set2*, and the conditions of *Set2* are weaker compared to those of *Set1*.

### C. Multiple access-cognitive interference network (MA-CIFN)

Now, we extend the result of Theorem 6 to a network with *k* + 1 transmitters and two receivers; a *k*-user MAC as a primary network and a point-to-point channel with a cognitive transmitter. We call it Multiple Access-Cognitive Interference Network (MA-CIFN). Consider MA-CIFN in Figure 3, denoted by \left({\mathcal{X}}_{1}\times {\mathcal{X}}_{2}\times \cdot \cdot \cdot \times {\mathcal{X}}_{k}\times {\mathcal{X}}_{k+1},p\left({y}_{1}^{n},{y}_{k+1}^{n}|{x}_{1}^{n},{x}_{2}^{n},\cdot \cdot \cdot {x}_{k}^{n},{x}_{k+1}^{n}\right),{\mathcal{Y}}_{1}\times {\mathcal{Y}}_{k+1}\right), where {X}_{j}\in {\mathcal{X}}_{j} is the channel input at Transmitter *j* (Tx*j*), for j\in \left\{1,...,k+1\right\};{Y}_{1}\in {\mathcal{Y}}_{1} and {Y}_{k+1}\in {\mathcal{Y}}_{k+1} are channel outputs at the primary and cognitive receivers, respectively, and p\left({y}_{1}^{n},{y}_{k+1}^{n}|{n}_{1}^{n},{x}_{2}^{n},...,{x}_{k}^{n},{x}_{k+1}^{n}\right) is the channel transition probability distribution. In *n* channel uses, each Tx*j* desires to send a message pair *m*_{
j
}to the primary receiver where *j* ∈ {1, ..., *k*}, and Tx*k* + 1 desires to send a message *m*_{k+1}to the cognitive receiver. We ignore the common information for brevity. Definitions 1 and 2 can be simply extended to the MA-CIFN. Therefore, we state the result on the capacity region under strong interference conditions.

*Corollary 1:* The capacity region of the MA-CIFN, satisfying

for all *S* ⊆ [1: *k*] and for every *p*(*x*_{1})*p*(*x*_{2})*...p*(*x*_{
k
}) *p*(*x*_{k+ 1}*|x*_{1},*x*_{2},...,*x*_{
k
})*p*(*y*_{1},*y*_{k+1}|*x*_{1},*x*_{2},...,*x*_{
k
},*x*_{k+1}), is given by

for all *S* ⊆ [1: *k*], where *X*(*S*) is the ordered vector of *X*_{
j
}, *j* ∈ *S*, and *S*^{c}denotes the complement of the set *S.*

*Proof:* Following the same lines as the proof of Theorem 6, the proof is straightforward. Therefore, it is omitted for the sake of brevity.

*Remark 8:* Under condition (51), the message of the cognitive user (*m*_{k+1}) can be decoded at the primary receiver (*Y*_{1}). Also, the cognitive receiver (*Y*_{k+1}), under condition (52), can decode *m*_{
j
}; *j* ∈ {1,...,*k*} in a MAC fashion. Therefore, the bound in (53) gives the capacity of a point-to-point channel with message *m*_{k+1}with side-information *X*_{
j
}; *j* ∈ {1,..., *k*} at the cognitive receiver. Moreover, (53) and (54) with condition (51), give the capacity region for a *k +* 1-user MAC with common information at the primary receiver.

## 6. Gaussian MA-CIFC

In this section, we consider the Gaussian MA-CIFC and characterize capacity results for the Gaussian case in the *weak* and *strong* interference regimes. For simplicity, we assume that Tx1 and Tx2 have no common information. This means that, *R*_{0} = 0 and {\mathcal{M}}_{0}=\varnothing. to investigate these regions. The Gaussian MA-CIFC, as depicted in Figure 4, at time *i* = 1,..., *n* can be mathematically modeled as

where *h*_{31}, *h*_{13}, and *h*_{23} are known channel gains. *X*_{1,i}, *X*_{2,i} and *X*_{3,i}are input signals with average power constraints:

for *j* ∈ {1,2, 3}. *Z*_{1,i}and *Z*_{3,i}are independent and identically distributed (i.i.d) zero mean Gaussian noise components with unit powers, i.e., {Z}_{j,i}~\mathcal{N}\left(0,1\right) for *j* ∈ {1, 3}.

### A. Strong interference regime

Here, we extend the results of Theorem 6, i.e., {\mathcal{C}}_{1}^{\mathsf{\text{str}}} and *Set1*, to the Gaussian case. The strong interference conditions of *Set1*, i.e., (34)-(37), for the above Gaussian model become:

where - 1 ≤ *ρ*_{
u
}≤ 1 is the correlation coefficient between *X*_{
u
}and *X*_{3}, i.e., E\left({X}_{u},{X}_{3}\right)={\rho}_{u}\sqrt{{P}_{u}{P}_{3}} for *u* ∈ {1, 2}.

*Theorem 8:* For the Gaussian MA-CIFC satisfying conditions (58)-(61), the capacity region is given by

where, to simplify notation, we define

*Remark 9:* Condition (58) implies that Tx3 causes strong interference at Rx1. This enables Rx1 to decode *m*_{3}. Moreover, (59)-(61) provide strong interference conditions at Rx3, under which all messages can be decoded in Rx3 in a MAC fashion.

*Proof:* The achievability part follows from {\mathcal{C}}_{1}^{\mathsf{\text{str}}} in Theorem 6 by evaluating (38)-(41) with zero mean jointly Gaussian channel inputs *X*_{1}, *X*_{2} and *X*_{3}. That is, {X}_{1}~\mathcal{N}\left(0,{P}_{1}\right),{X}_{2}~\mathcal{N}\left(0,{P}_{2}\right), and {X}_{3}~\mathcal{N}\left(0,{P}_{3}\right), where *E*(*X*_{1},*X*_{2}) = 0, E\left({X}_{1},{X}_{3}\right)={\rho}_{1}\sqrt{{P}_{1}{P}_{3}}, and E\left({X}_{2},{X}_{3}\right)={\rho}_{2}\sqrt{{P}_{2}{P}_{3}}. The converse proof is based on reasoning similar to that in [26] and is provided in "Appendix C".

It is noted that the channel parameters, i.e., *P*_{1},*P*_{2},*P*_{3},*h*_{31},*h*_{13},*h*_{23}, must satisfy (58)-(61) for all -1\le {\rho}_{1},{\rho}_{2}\le 1:{\rho}_{1}^{2}+{\rho}_{2}^{2}\le 1, to numerically evaluate the {\mathcal{C}}_{1}^{G} using (62)-(65). Here, we choose {P}_{1}={P}_{2}={P}_{3}=6,{h}_{31}={h}_{13}={h}_{23}=\sqrt{1.5} which satisfy strong interference conditions (58)-(61); hence, the regions are derived under strong interference conditions.

Figure 5 shows the capacity region for the Gaussian MA-CIFC of Theorem 8, for *P*_{1} = *P*_{2} = *P*_{3} = 6, and {h}_{31}={h}_{13}={h}_{23}=\sqrt{1.5}, where ρ_{1} = *ρ*_{2} is fixed in each surface. The *ρ*_{1} = *ρ*_{2} = 0 region corresponds to the *no cooperation* case, where the channel inputs are independent. It can be seen that as *ρ*_{1} = *ρ*_{2} increases, the bound on *R*_{3} becomes more restrictive while the sum-rate bounds become looser; because Tx3 dedicates parts of its power for cooperation. This means that, as Tx3 allocates more power to relay *m*_{1}, *m*_{2} by increasing *ρ*_{1} = *ρ*_{2}, *R*_{1} and *R*_{2} improve, while *R*_{3} degrades due to the less power allocated to transmit *m*_{3}. The capacity region for this channel is the union of all the regions obtained for different values of *ρ*_{1} and *ρ*_{2}, satisfying {\rho}_{1}^{2}+{\rho}_{2}^{2}\le 1. This union is shown in Figure 6. In order to better perceive the effect of cooperation, we let *R*_{2} = 0 in Figure 7. It is seen that by increasing *ρ*_{1} = *ρ*_{2}, the bound on *R*_{1} + *R*_{3} becomes looser and *R*_{1} improves, while *R*_{3} decreases due to the more power dedicated for cooperation.

### B. Weak interference regime

Now, we consider the Gaussian MA-CIFC with weak interference at the primary receiver (Rx1), which means *h*_{31} ≤ 1. We remark that, since there is no cooperation between receivers, the capacity region for this channel is the same as the one with the same marginal outputs p\left({y}_{1}^{n}|{x}_{1}^{n},{x}_{2}^{n},{x}_{3}^{n}\right) and p\left({y}_{3}^{n}|{x}_{1}^{n},{x}_{2}^{n},{x}_{3}^{n}\right). Hence, we can state the following useful lemma.

*Lemma 2:* The capacity region of a Gaussian MA-CIFC, defined by (55) and (56) when *h*_{31} ≤ 1, is the same as the capacity region of a Gaussian MA-CIFC with the following channel outputs:

where *Y'*_{3,i}= *X*_{3,i}+ *Z*_{3,i}and {Z}_{1,i}^{\prime}~\mathcal{N}\left(0,1-{h}_{31}^{2}\right). Therefore, the degradedness condition in (29) holds for the Gaussian MA-CIFC when *h*_{31} ≤ 1.

*Proof:* The proof follows from [6, Lemma 3.5].

Next, we use the inner bound of Theorem 2 and the outer bound of Theorem 4 to derive the capacity region, which shows that the capacity-achieving scheme in this case consists of DPC at the cognitive transmitter and treating interference as noise at both receivers.

*Theorem 9:* For the Gaussian MA-CIFC, defined by (55) and (56), when *h*_{31} ≤ 1, the capacity region is given by

where *θ*(·) is defined in (66).

*Remark 10:* By evaluating (2) with jointly Gaussian channel inputs, one can easily achieve (69). However, this results in the Gaussian counterparts of the bounds in (7)-(10). Therefore, some conditions are necessary to make these bounds redundant, similar to the ones in (30)-(33). However, we show that (12) is also evaluated to (69), if we apply DPC with appropriate parameters. Hence, conditions (30)-(33) are unnecessary in the Gaussian case. This means that DPC completely mitigates the effects of interference for the Tx3-Rx3 pair and leaves the link between them interference-free for fixed values of *ρ*_{1}, *ρ*_{2}. Consequently, {\mathcal{C}}_{2}^{G} is independent of *h*_{13} and *h*_{23}.

*Remark 11:* If we omit Tx2, the model reduces to a CIFC and by setting {P}_{2}={\rho}_{2}={R}_{2}=0,{\mathcal{C}}_{2}^{G} coincides with the capacity region of the Gaussian CIFC with weak interference, which was characterized in [6, Lemma 3.6].

*Proof:* The rate region in Theorem 2 can be extended to the discrete-time Gaussian memoryless case with continuous alphabets by standard arguments [23]. Hence, it is sufficient to evaluate (3)-(6) and (12) with an appropriate choice of input distribution to reach (69)-(72). Let {R}_{0}=0,{\mathcal{M}}_{0}=\varnothing, and T=\varnothing, since Tx1 and Tx2 have no common information. Also, let *U* and *V* be deterministic constants. We choose zero mean jointly Gaussian channel inputs *X*_{1}, *X*_{2} and *X*_{3}. In fact, {X}_{j}~\mathcal{N}\left(0,{P}_{j}\right) for *j* ∈ {1,2,3}, where *E*(*X*_{1},*X*_{2}) = 0, E\left({X}_{1},{X}_{3}\right)={\rho}_{1}\sqrt{{P}_{1}{P}_{3}}, and E\left({X}_{2},{X}_{3}\right)={\rho}_{2}\sqrt{{P}_{2}{P}_{3}}. Noting the p.m.f (13), consider the following choice of input distribution for certain \left\{-1\le {\rho}_{1},{\rho}_{2}\le 1:{\rho}_{1}^{2}+{\rho}_{2}^{2}\le 1\right\}:

Therefore, (3)-(6) are easily evaluated to (70)-(72). In "Appendix D", we derive (69) by evaluating (12) with appropriate parameters. The converse proof follows by applying the bounds in the proof of Theorem 4 to the Gaussian case and utilizing Entropy Power Inequality (EPI) [23, 27]. A detailed converse proof is provided in "Appendix D".

*Remark 12:* According to Theorems 8 and 9, jointly Gaussian channel inputs *X*_{1}, *X*_{2} and *X*_{3} are optimal for the Gaussian MA-CIFC under the strong and weak interference conditions, determined in the above theorems.

Figure 8 shows the capacity region for the Gaussian MA-CIFC of Theorem 9, for *P*_{1} = *P*_{2} = *P*_{3} = 6, and {h}_{31}=\sqrt{0.55}, where *ρ*_{1} = *ρ*_{
2
}is fixed in each surface. It is noted that the capacity region is independent of *h*_{13} and *h*_{23}. The *ρ*_{1} = *ρ*_{2} = 0 region corresponds to the *no cooperation* case, where channel inputs are independent. We see that when Tx3 dedicates parts of its power for cooperation, i.e., *ρ*_{1} = *ρ*_{2} = 0.5, the rates of the primary users (*R*_{1}, R_{2}) increase, while *R*_{3} decreases. The capacity region for this channel is the union of all the regions obtained for different values of *ρ*_{1} and *ρ*_{2} satisfying, {\rho}_{1}^{2}+{\rho}_{2}^{2}\le 1, which is shown in Figure 9. Similar to Figure 7, we investigate the capacity region for *R*_{2} = 0 in Figure 10 in the weak interference regime. It is seen that, when Tx3 dedicates more power for cooperation by increasing *ρ*_{1} = *ρ*_{2}, *R*_{1} improves and *R*_{3} decreases.

## 7. Conclusion

We investigated a cognitive communication network where a MAC with common information and a point-to-point channel share a same medium and interfere with each other. For this purpose, we introduced Multiple Access-Cognitive Interference Channel (MA-CIFC) by merging a two-user MAC as a primary network and a cognitive transmitter-receiver pair in which the cognitive transmitter knows the message being sent by all of the transmitters in a non-causal manner. We analyzed the capacity region of MA-CIFC by deriving the inner and outer bounds on the capacity region of this channel. These bounds were proved to be tight in some special cases. Therefore, we determined the optimal strategy in these cases. Specifically, in the discrete memoryless case, we established the capacity regions for a class of *degraded* MA-CIFC and also under two sets of strong interference conditions. We also derive strong interference conditions for a network with *k* primary users. Further, we characterized the capacity region of the Gaussian MA-CIFC in the *weak* and *strong* interference regimes. We showed that DPC at the cognitive transmitter and treating interference as noise at the receivers, i.e., an oblivious primary receiver, are optimal in the weak interference. However, the receivers have to decode all messages when the interference is strong enough, which requires an aware primary receiver.

## Appendix A Proofs of Theorems 1, 2 and 3

*Outline of the proof for Theorem 1:* We propose the following random coding scheme, which contains superposition coding and the technique of [6] in defining auxiliary RVs, i.e., *U* and *V*. The cognitive receiver (Rx3) decodes the interfering signals caused by the primary messages (*m*_{1}, *m*_{2}), while the primary receiver (Rx1) does not decode the interference from the cognitive user's message (*m*_{3}) and treats it as noise.

** Codebook Generation**: Fix a joint p.m.f as (11). Generate {2}^{n{R}_{0}} i.i.d

*t*

^{n}sequences, according to the probability \prod _{i=1}^{n}p\left({t}_{i}\right). Index them as t

^{n}(

*m*

_{0}) where {m}_{0}\in \left[1,{2}^{n{R}_{0}}\right]. For each t

^{n}(

*m*

_{0}), generate {2}^{n{R}_{1}} i.i.d \left({u}^{n},{x}_{1}^{n}\right) sequences, each with probability \prod _{i=1}^{n}p\left({u}_{i},{x}_{1,i}|{t}_{i}\right). Index them as \left({u}^{n}\left({m}_{0},{m}_{1}\right),{x}_{1}^{n}\left({m}_{0},{m}_{1}\right)\right) where {m}_{1}\in \left[1,{2}^{n{R}_{1}}\right]. Similarly, for each t

^{n}(

*m*

_{0}), generate {2}^{n{R}_{2}} i.i.d \left({v}^{n},{x}_{2}^{n}\right) sequences, each with probability \prod _{i=1}^{n}p\left({v}_{i},{x}_{2,i}|{t}_{i}\right). Index them as \left({v}^{n}\left({m}_{0},{m}_{2}\right),{x}_{2}^{n}\left({m}_{0},{m}_{2}\right)\right) where {m}_{2}\in \left[1,{2}^{n{R}_{2}}\right]. For each \left({t}^{n}\left({m}_{0}\right),{u}^{n}\left({m}_{0},{m}_{1}\right),{x}_{1}^{n}\left({m}_{0},{m}_{1}\right),{v}^{n}\left({m}_{0},{m}_{2}\right),{x}_{2}^{n}\left({m}_{0},{m}_{2}\right)\right), generate {2}^{n{R}_{3}} i.i.d {x}_{3}^{n} sequences, according to \prod _{i=1}^{n}p\left({x}_{3,i}|{t}_{i},{u}_{i},{x}_{1,i},{v}_{i},{x}_{2,i}\right). Index them as {x}_{3}^{n}\left({m}_{0},{m}_{1},{m}_{2},{m}_{3}\right) where {m}_{3}\in \left[1,{2}^{n{R}_{3}}\right].

** Encoding**: In order to transmit the messages (

*m*

_{0},

*m*

_{1},

*m*

_{2},

*m*

_{3}), Tx

*j*sends {x}_{j}^{n}\left({m}_{0},{m}_{j}\right) for

*j*∈ {1,2} and Tx3 sends {x}_{3}^{n}\left({m}_{0},{m}_{1},{m}_{2},{m}_{3}\right).

*Decoding:*

*Rx1:* After receiving {y}_{1}^{n}, Rx1 looks for a unique triple \left({\widehat{m}}_{0},{\widehat{m}}_{1},{\widehat{m}}_{2}\right) and some {\widehat{m}}_{2} such that

For large enough *n* with arbitrarily high probability, \left({\widehat{m}}_{0},{\widehat{m}}_{1},{\widehat{m}}_{2}\right)=\left({m}_{0},{m}_{1},{m}_{2}\right) if (3)-(6) hold.

*Rx3:* After receiving {y}_{3}^{n}, Rx3 finds a unique index {\widehat{\widehat{m}}}_{3} and some triple \left({\widehat{\widehat{m}}}_{0},{\widehat{\widehat{m}}}_{1},{\widehat{\widehat{m}}}_{2}\right) such that

With arbitrary high probability, {\widehat{\widehat{m}}}_{3}={m}_{3} if *n* is large enough and (2), (7)-(10) hold. This completes the proof.

*Outline of the proof for Theorem 2:* Our proposed random coding scheme, in the encoding part, contains the methods of Theorem 1 and GP binning at the cognitive transmitter (Tx3) which is used at Tx3 to cancel the interference caused by *m*_{0}, *m*_{1}, *m*_{2} at Rx3. In the decoding part, both receivers decode only their intended messages, treating the interference as noise. Therefore, unlike the decoding part of Theorem 1, Rx3 decodes only its own message (*m* 3), while treating the other signals as noise.

** Codebook Generation**: Fix a joint p.m.f as (13). Generate {t}^{n}\left({m}_{0}\right),{u}^{n}\left({m}_{0},{m}_{1}\right),{x}_{1}^{n}\left({m}_{0},{m}_{1}\right),{v}^{n}\left({m}_{0},{m}_{2}\right),{x}_{2}^{n}\left({m}_{0},{m}_{2}\right) codewords based on the same lines as in the

*codebook generation*part of Theorem 1. Then, generate {2}^{n\left({R}_{3}+L\right)} i.i.d

*w*

^{n}sequences. Index them as

*w*

^{n}(

*m*

_{3},

*l*) where {m}_{3}\in \left[1,{2}^{n{R}_{3}}\right] and

*l*∈ [1,2

^{nL}].

** Encoding**: In order to transmit the messages (

*m*

_{0},

*m*

_{1},

*m*

_{2},

*m*

_{3}), Tx

*j*sends {x}_{j}^{n}\left({m}_{0},{m}_{j}\right) for

*j*∈ {1,2}. Tx3 (the cognitive transmitter), in addition to

*m*

_{3}, knows

*m*

_{0},

*m*

_{1}and

*m*

_{2}. Hence, knowing codewords {t}^{n},{u}^{n},{x}_{1}^{n},{v}^{n},{x}_{2}^{n}, to send

*m*

_{3}, it seeks an index

*l*such that

If there is more than one such index, Tx3 picks the smallest. If there is no such codeword, it declares an error. Using covering lemma [27], it can be shown that there exists such an index *l* with high enough probability if *n* is sufficiently large and

Then, Tx3 sends {x}_{3}^{n} generated according to \prod _{i=1}^{n}p\left({x}_{3,i}|{t}_{i},{u}_{i},{x}_{1,i},{v}_{i},{x}_{2,i}\right).

** Decoding**: The decoding procedure at Rx1 is similar to Theorem 1 and the error in this receiver can be bounded if (3)-(6) hold.

*Rx3:* After receiving {y}_{3}^{n}, Rx3 finds a unique index {\widehat{\widehat{m}}}_{3} for some index \widehat{\widehat{l}} such that

For large enough *n*, the probability of error can be made sufficiently small if

Combining (74) and (75) results in (12). This completes the proof.

*Outline of the proof for Theorem 3:* We propose the following random coding scheme, which contains superposition coding in the encoding part and simultaneous joint decoding in the decoding part. All messages are common to both receivers, i.e., both receivers decode (*m*_{0}, *m*_{1}, *m*_{2}, *m*_{3}).

** Codebook Generation**: Fix a joint p.m.f as (18). Generate {2}^{n{R}_{0}} i.i.d

*t*

^{n}sequences, according to the probability \prod _{i=1}^{n}p\left({t}_{i}\right). Index them as

*t*

^{n}(

*m*

_{0}) where {m}_{0}\in \left[1,{2}^{n{R}_{0}}\right]. For

*j*∈ {1,2} and each

*t*

^{n}(

*m*

_{0}), generate {2}^{n{R}_{j}} i.i.d {x}_{j}^{n} sequences, each with probability \prod _{i=1}^{n}p\left({x}_{j,i}|{t}_{i}\right). Index them as {x}_{j}^{n}\left({m}_{0},{m}_{j}\right) where {m}_{j}\in \left[1,{2}^{n{R}_{j}}\right]. For each \left({t}^{n}\left({m}_{0}\right),{x}_{1}^{n}\left({m}_{0},{m}_{1}\right),{x}_{2}^{n}\left({m}_{0},{m}_{2}\right)\right), generate {2}^{n{R}_{3}} i.i.d {x}_{3}^{n} sequences, each with probability \prod _{i=1}^{n}p\left({x}_{3,i}|{t}_{i},{x}_{1,i},{x}_{2,i}\right). Index them as {x}_{3}^{n}\left({m}_{0},{m}_{1},{m}_{2},{m}_{3}\right) where {m}_{3}\in \left[1,{2}^{n{R}_{3}}\right].

** Encoding**: In order to transmit message (

*m*

_{0},

*m*

_{1},

*m*

_{2},

*m*

_{3}), Tx

*j*sends {x}_{j}^{n}\left({m}_{0},{m}_{j}\right) for

*j*∈ {1,2} and Tx3 sends {x}_{3}^{n}\left({m}_{0},{m}_{1},{m}_{2},{m}_{3}\right).

*Decoding:*

*Rx1:* After receiving {y}_{1}^{n}, Rx1 looks for a unique triple \left({\widehat{m}}_{0},{\widehat{m}}_{1},{\widehat{m}}_{2}\right) and some {\widehat{m}}_{3} such that

For large enough *n*, with arbitrarily high probability \left({\widehat{m}}_{0},{\widehat{m}}_{1},{\widehat{m}}_{2}\right)=\left({m}_{0},{m}_{1},{m}_{2}\right) if

*Rx3:* Similarly, after receiving {y}_{3}^{n}, Rx3 finds a unique index {\widehat{\widehat{m}}}_{3} and some triple \left({\widehat{\widehat{m}}}_{0},{\widehat{\widehat{m}}}_{1},{\widehat{\widehat{m}}}_{2}\right) such that

With the arbitrary high probability {\widehat{\widehat{m}}}_{3}={m}_{3}, if *n* is large enough and