Cooperative Relaying with State Available at the Relay

# Cooperative Relaying with State Available at the Relay

Abdellatif Zaidi   Shivaprasad Kotagiri   J. Nicholas Laneman   Luc Vandendorpe
École Polytechnique, Université Catholique de Louvain, LLN-1348, Belgium
Department of Electrical Engineering, University of Notre, Notre Dame, IN-46556, USA
{zaidi,vandendorpe}@uclouvain.be, {skotagir,jnl}@nd.edu
The work by A. Zaidi and L. Vandendorpe was supported in part by the EU framework program COOPCOM and the Walloon region framework program NANOTIC-COSMOS. The work by S. Kotagiri and J. Nicholas Laneman was supported in part by the State of Indiana through the Twenty-First Century Research and Technology Fund.
###### Abstract

We consider a state-dependent full-duplex relay channel with the state of the channel non-causally available at only the relay. In the framework of cooperative wireless networks, some specific terminals can be equipped with cognition capabilities, i.e, the relay in our model. In the discrete memoryless (DM) case, we derive lower and upper bounds on channel capacity. The lower bound is obtained by a coding scheme at the relay that consists in a combination of codeword splitting, Gel’fand-Pinsker binning, and a decode-and-forward scheme. The upper bound is better than that obtained by assuming that the channel state is available at the source and the destination as well. For the Gaussian case, we also derive lower and upper bounds on channel capacity. The lower bound is obtained by a coding scheme which is based on a combination of codeword splitting and Generalized dirty paper coding. The upper bound is also better than that obtained by assuming that the channel state is available at the source, the relay, and the destination. The two bounds meet, and so give the capacity, in some special cases for the degraded Gaussian case.

## I Introduction

Channels that depend on random parameters have received considerable attention over the last decade, due to a wide range of possible applications. This includes single user models [1, 2, 3, 4] and multiuser models (see, e.g., [5, 6, 7] and references therein). For multiuser models, one key issue in the study of state-dependent channels is whether the parameters controlling the channel are known symmetrically, i.e., to all, or asymmetrically, i.e., to only some of, the users in the communication model. The broadcast channel (BC) with state available at the transmitter but not at the receivers is considered in [5, 6, 8]. The multiple access channel (MAC) with partial state information at all the encoders and full state information at the decoder is considered in [9].

In the Gaussian case, the MAC with all informed encoders, the BC with informed encoder, the physically degraded relay channel (RC) with informed source and informed relay, and the physically degraded relay broadcast channel (RBC) with informed source and informed relay are studied in [5, 10, 11]. In all these cases, it is shown that some variants of Costa’s dirty paper coding (DPC) [3] achieve the respective capacity or the respective capacity region. Also, since for all these models the variant of DPC achieves the trivial upper or outer bound obtained by assuming that the channel state is also available at the decoders in the model, it is not required to obtain any non-trivial upper or outer bounds. For all these models, the key assumption that makes the problem relatively easy is the availability of the channel state at all the encoders in the communication model. It is interesting to study state-dependent multi-user models in which only some, i.e., not all, the encoders are informed about the channel state, because the uninformed encoders in the model cannot apply DPC.

The state-dependent MAC with some, but not all, encoders informed of the channel state is considered in [12, 13, 14, 15, 16] and the state-dependent relay channel with only informed source is considered in [11, 17]. For all these models, in the Gaussian case, the informed encoder applies a slightly generalized DPC (GDPC) in which the channel input and the channel state are correlated. In these models, the uninformed encoders benefit from the GDPC applied by the informed encoders because the negative correlation between the codewords at the informed encoders and the channel state can be interpreted as partial state cancellation. For the state-dependent MAC with one informed encoder, the capacity region for the Gaussian case is obtained by deriving a non-trivial upper bound in the case in which the message sets are degraded [15].

In this work, we consider a three terminal state-dependent full-duplex RC in which the state of the channel is known non-causally to only the relay, i.e., but neither to the source nor to the destination. We refer to this communication model as state-dependent RC with informed relay. This model is shown in Figure 1.

For the discrete memoryless (DM) case, we derive a lower bound (Section III) and an upper bound (Section IV) on the capacity of the state-dependent general, i.e., not necessarily degraded, RC with informed relay. The lower bound is obtained by a coding scheme at the relay that uses a combination of codeword splitting, Gel’fand-Pinsker coding, and regular encoding backward decoding [18] for full-duplex decode-and-forward (DF) relaying [19]. The upper bound on the capacity is better than that obtained by assuming that the channel state is also available at the source and the destination. Also, this upper bound is non-trivial and connects with a bounding technique which is developed in the context of multiple access channels with asymmetric channel state in [16, Theorem 2]. However, we note that, in this paper, the upper bound is proved using techniques that are different from those in [16]. Furthermore, we also specialize the results in the DM case to the case in which the channel is physically degraded.

For the Gaussian case (Section V), we derive lower and upper bounds on channel capacity by applying the concepts developed in the DM case to the case in which the CSI is additive Gaussian, i.e., models an additive Gaussian interference, and the ambient noise is additive Gaussian. Furthermore, we point out the loss caused by the asymmetry and show that the lower bound is (in general) close and is tight in a number of special cases if the channel is physically degraded. The key idea for the lower bound is an appropriate code construction which allows the source and the relay to send coherent signals (by enabling correlation between source and relay signals, though only one of the two encoders is informed) and, at the same time, have the source possibly benefit from the availability of the CSI at the relay (through a generalized DPC). Also, we characterize the optimal strategy of the relay balancing the trade-off between enhancing source’s transmitted signal and combating the interference. Section VI presents some illustrative numerical examples.

## Ii The DM RC With Informed Relay Only

The state-dependent DM RC, denoted by and depicted in Figure 1, consists of five finite sets , a distribution on the set and a family of conditional probability distributions from to . Let and designate the inputs of the source and the relay respectively, and and designate the outputs of the relay and the channel, respectively. We assume that the channel states are i.i.d., each distributed according to .

As it can be seen from Figure 1, the setup we consider is asymmetric in the sense that only the relay is informed of the channel states. The relay observes the CSI non-causally, and we refer to this channel as state-dependent RC with informed relay. Also, the channel will be said to be physically degraded if the conditional distribution can be written as

 WY2,Y3|X1,X2,S=WY2|X1,X2,SWY3|Y2,X2,S. (1)

In Section III and Section IV we establish bounds on the capacity of the state-dependent RC with informed relay in the DM case. We assume that the alphabets are finite. The proofs of these bounds are rather lengthy and, for the sake of brevity, they are either outlined only or omitted here. Detailed proofs are reported in [20].

## Iii DM Case: Lower Bound on Capacity

The following theorem provides a lower bound on the capacity of the state-dependent DM RC with informed relay.

###### Theorem 1

The capacity of the state-dependent DM RC with informed relay is lower bounded by

 Rin=maxmin{ I(X1;Y2|S,U1), I(X1,U1,U2;Y3)−I(U2;S|U1)}, (2)

where the maximization is over all joint measures of the form

 PS,U1,U2,X1,X2,Y2,Y3= QSPU1PX1|U1PU2|U1,SPX2|U1,U2,SWY2,Y3|X1,X2,S (3)

and and are auxiliary random variables with cardinality bounded as

 |U1|≤|S||X1||X2|+1 (4a) |U2|≤(|S||X1||X2|+1)|S||X1||X2|, (4b)

respectively.

The proof of Theorem 1 is based on a random coding scheme the basic idea of which is conveyed in the following remarks.

###### Remark 1

The lower bound (2) is based on the relay operating in a decode-and-forward (DF) relaying scheme [19, Theorem 1]. In DF strategies, the source knows what cooperative information the relay transmits. Consequently, the source generates the codeword that achieves coherence gains as in multi-antenna transmission, by having the channel input of the source correlated with that of the relay. In our model, at one hand, if the relay generates its channel input as a function of both the cooperative information and the channel state, then it is not possible for the source to know the relay input, because the source does not know the channel state. On the other hand, the relay should not completely ignore the known channel state which it can use to remove the effect of the CSI on the communication. To resolve the existing tension at the relay, we generate two codebooks. In one codebook, the random codewords are generated using a random variable which is independent of the channel state . The relay chooses the appropriate random codeword from this codebook using the cooperative information only. In another codebook, the codewords are generated using the random variable which is correlated with the channel state and through . The relay chooses the codeword from this codebook using both the cooperative information and the channel state, to remove the effect of the channel state on the communication. Then the relay generates the channel input from using the conditional probability law . Likewise, the source knows because it is a function of only the cooperative information, and, given , it generates the random codeword according to the conditional probability law . Thus the channel inputs of the source and the relay are correlated through . This is the main idea of the coding scheme that we use at the relay.

###### Remark 2

The term in (2) illustrates the multi-antenna system behavior when the encoders are asymmetrically informed about the channel: it can be interpreted as an achievable sum rate over a two-encoder MAC with only one encoder being informed of the CSI. In [15, 14], the authors derive the capacity region of a state-dependent MAC with one informed encoder in the case in which the informed encoder knows the message to be transmitted by the uninformed encoder. In our model, transmission from the source and the relay to the destination can also be viewed as that over a MAC with one informed encoder. However, in our case, due to the structure of the DF scheme, the situation is different from [15, 14] since the uninformed encoder (the source) knows the message of the informed encoder (the relay), not the opposite. This makes coding at the relay more difficult in our setup.

## Iv DM Case: Upper Bound on Capacity

The following theorem provides an upper bound on the capacity of the state-dependent DM RC with informed relay.

###### Theorem 2

The capacity of the state-dependent DM RC with informed relay is upper bounded by

 Rout=maxmin{ I(X1;Y2,Y3|S,X2), I(X1,X2;Y3|S)−I(X1;S|Y3)}, (5)

where the maximization is over all joint measures of the form

 PS,X1,,X2,Y2,Y3=QSPX1PX2|X1,SWY2,Y3|X1,X2,S. (6)

In the second term of the minimization in (5), the term can be interpreted as the rate penalty in the information conveyed to the destination caused by not knowing the channel state at the source as well. This rate loss makes the above upper bound tighter than the trivial upper bound which is obtained by assuming that the channel state is also available at the source and the destination, i.e.,

 Routtriv=maxmin{I(X1;Y2,Y3|S,X2),I(X1,X2;Y3|S)}, (7)

where here maximization is over all joint measures of the form

 PS,X1,,X2,Y2,Y3=QSPX1|SPX2|X1,SWY2,Y3|X1,X2,S. (8)

If the channel is physically degraded, the upper bound in Theorem 2 reduces to the one in the following corollary.

###### Corollary 1

The capacity of the state-dependent physically degraded DM RC with informed relay is upper bounded by

 RoutD=maxmin{ I(X1;Y2|S,X2), I(X1,X2;Y3|S)−I(X1;S|Y3)} (9)

where the maximization is over all probability distributions of the form

 PS,X1,X2,Y2,Y3 =QSPX1PX2|X1,SWY2|X1,X2,SWY3|Y2,X2,S. (10)

## V The Gaussian Relay Channel

In this section, we consider a state-dependent full-duplex Gaussian RC in which both the channel state and the noise are additive and Gaussian. In this model also, we assume that the additive channel state is non-causally known to only the relay.

### V-a Channel Model

For our model of the state-dependent full-duplex Gaussian RC with informed relay, the channel outputs and at time instant for the relay and the destination, respectively, are related to the channel input from the source and from the relay, and the channel state by

 Y2,i =X1,i+Si+Z2,i, (11a) Y3,i =X1,i+X2,i+Si+Z3,i, (11b)

where is a zero mean Gaussian random variable with variance , is zero mean Gaussian with variance , and is zero mean Gaussian with variance . The random variables , and at time instant are mutually independent, are independent of the channel inputs , and are independent of for .

For the full-duplex degraded additive Gaussian RC, the channel outputs and for the relay and the destination, respectively, are related to the channel inputs and and the state by

 Y2,i =X1,i+Si+Z2,i, (12a) Y3,i =X2,i+Y2,i+Z′3,i, (12b)

where is a sequence of i.i.d. zero mean Gaussian random variables with variance which is independent of .

We consider individual power constraints on the transmitted power, .

### V-B Bounds on Capacity

The results obtained in Section IV for the DM case can be applied to memoryless channels with discrete time and continuous alphabets using standard techniques [21]. We use the bounds in Theorem 1 and Theorem 2 to compute bounds on channel capacity for the Gaussian case.

The following theorem provides a lower bound on the capacity of the state-dependent Gaussian RC with informed relay.

###### Theorem 3

The capacity of the state-dependent general Gaussian RC with informed relay is lower bounded by

 RinG=maxρ′12min{12log(1+P1(1−ρ′212)N2), +12log(1+θP2(1−ρ′22s)N3)}, (13)

where the maximization is over parameters , , , and .

The proof of Theorem 3 is based on the evaluation of the lower bound (2) with an appropriate jointly Gaussian input distribution that will be specified in the sequel.

Recalling the discussion in Remark 1, for the Gaussian RC with informed relay, we should consider two important features in the design of an efficient coding scheme at the relay: obtaining correlation or coherence between the channel inputs from the source and the relay, and exploiting the channel state to remove the effect of the CSI on the communication. As we already mentioned, it is not obvious to accomplish these features because the channel state is not available at the source. The main idea in the coding scheme that we consider consists in splitting the relay input into two independent parts, namely and . The first part, , is a function of only the cooperative information, and is generated using standard coding. Since the source knows the cooperative information at the relay, it can generate its codeword in such a way that it is coherent with , i.e., by allowing correlation between and . The second part, , which is independent of the source input , is a function of both the cooperative information and the channel state at the relay, and is generated using a GDPC similar to that in [11, 12, 13, 16, 22].

More formally, we decompose the relay input random variable as

 X2=U1+~X2, (14)

where: is zero mean Gaussian with variance , is independent of both and , and is correlated with with , for some , ; and is zero mean Gaussian with variance , is independent of , and is correlated with the channel state with , for some . Using the covariances and , the parameters , are given by

 ρ′12=σ12√¯θP1P2,ρ′2s=σ2s√θP2Q. (15)

For the GDPC, we choose the auxiliary random variable as

 U2=~X2+αoptS (16)

with

 αopt=θP2(1−ρ′22s)θP2(1−ρ′22s)+N3(1+ρ′2s√θP2Q)−ρ′2s√θP2Q. (17)

We now provide an upper bound on the capacity of the state-dependent Gaussian RC with informed relay.

###### Theorem 4

The capacity of the state-dependent general Gaussian RC with informed relay is upper bounded by

 12log(1+(√P1+ρ12√P2)2P2(1−ρ212−ρ22s)+(√Q+ρ2s√P2)2+N3) +12log(1+P2(1−ρ212−ρ22s)N3)}, (18)

where the maximization is over parameters and such that

 ρ212+ρ22s≤1. (19)

The proof of Theorem 4 follows by evaluating the upper bound (5) using an appropriate joint distribution of . It is based on showing that for the Gaussian channel (11), one can restrict attention to jointly Gaussian with , and . The allowable values for the covariances and are such that the covariance matrix of has a non-negative discriminant, i.e. . For , this gives (19).

Similarly to in Theorem 4, we obtain an upper bound on channel capacity for the degraded Gaussian case by evaluating the upper bound (9) in Corollary 1 using an appropriate jointly Gaussian distribution of .

###### Corollary 2

The capacity of the state-dependent degraded Gaussian RC with informed relay is upper bounded by (18) in which the first term of the minimization is replaced by

 12log(1+P1(1−ρ212−ρ22s)N2(1−ρ22s)). (20)

### V-C Special Cases Analysis

We note that comparing the bounds in Theorem 3 and Theorem 4 analytically can be tedious in the general case. In the following, we focus only on the physically degraded case. In this case, we show that the lower bound in (13) is tight for certain values of the channel statistics, and thus obtain the capacity expression for these cases.

In the following corollary we recast the upper bound (18) into an equivalent form by substituting and .

###### Corollary 3

For the state-dependent degraded Gaussian RC with informed relay, the upper bound (18) can be written as

 RoutDG=maxκmin{12log(1+P1(1−κ2)N2), maxρ12log(1+P2(1−κ2(1−ρ2)−ρ2)N3) +12log(1+P1+κ2(1−ρ2)P2+2κ√1−ρ2√P1P2P2(1−κ2(1−ρ2))+Q+2ρ√P2Q+N3)}, (21)

where the maximization is over parameters and .

Investigating the lower bound (13) and the upper bound (21), it can be shown that the lower bound for the degraded case is tight for certain values of , , , , . The following proposition provides some cases for which the lower bound is tight.

###### Proposition 1

For the physically degraded Gaussian RC channel, we have the following.
If , , , , satisfy

 N2≥maxζ∈[−1,0]P1N3(P2+Q+N3+2ζ√P2Q)P1N3+P2(1−ζ2)(P1+P2+Q+N3+2ζ√P2Q), (22)

then channel capacity is given by

 CDG=12log(1+P1N2), (23)

which is the same as the interference-free capacity, i.e., the capacity when the channel state is also known to the source or is not present in the model.

If the maximizing and in the upper bound in Theorem 4 are such that condition (19) is met with equality, i.e., , then the lower bound (13) is tight and gives capacity.

Extreme Cases
We now summarize the behavior of the capacity in some extreme cases.

1. For , i.e., no channel state at all in the model, capacity is given by

 CDG=max0≤β≤1min{ 12log(1+P1(1−β2)N2), 12log(1+P1+P2+2β√P1P2N3)}, (24)

which is the same as the capacity of the standard degraded Gaussian channel [19, Theorem 5]. This can be directly obtained by substituting in (13) and (21). In this case, the maximizing parameters are , for (13) and for (21).

2. In the case of arbitrary strong channel state, i.e., , capacity is given by

 CDG=min{12log(1+P1N2),12log(1+P2N3)}. (25)

In this case, the lower bound in (13) is maximized for , and the upper bound in (21) is maximized for , and the two bounds meet. We note that, in this strong channel state case, (25) suggests that traditional multi-hop transmission achieves the capacity. A multi-hop scheme allows to completely cancel the effect of the channel state by subtracting it out upon reception at the relay, and by applying standard DPC for transmission from the relay to the destination.

3. If , capacity is given by

 CDG=12log(1+P1Q+N3). (26)

In this case, the informed relay cannot help the source, and the channel state is simply treated as an unknown noise.

## Vi Numerical Examples and Concluding Remarks

In this section we discuss some numerical examples, for both the degraded Gaussian case and the general Gaussian case.

Figure 2 illustrates the lower bound (13) and the upper bound (21) as functions of the signal-to-noise-ratio (SNR) at the relay, i.e., (in decibels). Also shown for comparison are the trivial upper bound (7) computed for the degraded Gaussian case and the trivial lower bound obtained by considering the channel state as an unknown noise. The curves show that the lower bound and the upper bound do not meet for all SNR regimes. However, as it is visible from the depicted numerical examples, the gap between the two bounds is small for the degraded case. Furthermore, the curves in Figure 2 also illustrate the results in proposition 1, by showing that the lower bound and the upper bound meet for the cases stated in Proposition 1. We note that the pentagram marker visible in Figure  2 indicates capacity when the noise at the relay is equal to the RHS of (22).

Figure 3 illustrates the lower bound (13) and the upper bound (18) as functions of the SNR at the relay for the general Gaussian channel. Also shown for comparison are the trivial upper bound (7) computed for the general Gaussian case and the trivial lower bound obtained by considering the channel state as an unknown noise. The curves show that the lower bound is close to the upper bound at large SNR, i.e., when capacity of the channel is determined by the sum rate of the MAC formed by transmission from the uninformed source and the informed relay to the destination. Furthermore, Figure 3 also shows the variation of the maximizing , , in (13) as function of the SNR at the relay. This shows how the informed relay allocates its power among combating the interference for the source (related to the value of ) and sending signals that are coherent with the transmission from the source (related to the values of and ).

## References

• [1] C. E. Shannon, “Channels with side information at the transmitter,” IBM journal of Research and Development, vol. 2, pp. 289–293, October 1958.
• [2] S. I. Gel’fand and M. S. Pinsker, “Coding for channel with random parameters,” Problems of Control and Information Theory, vol. 9, pp. 19–31, 1980.
• [3] M. H. M. Costa, “Writing on dirty paper,” IEEE Trans. on Info. Theory, vol. 29, pp. 439–441, May 1983.
• [4] C. D. Heegard and A. El Gamal, “On the capacity of computer memory with defects,” IEEE Trans. on Info. Theory, vol. IT-29, pp. 731–739, September 1983.
• [5] S. Gel’fand and M. Pinsker, “On Gaussian channel with random parameters,” in Proc. IEEE Int. Symp. on Info. Theory, Tashkent, 1984.
• [6] Y. Steinberg, “Coding for the degraded broadcast channel with random parameters, with causal and noncausal side information,” IEEE Trans. on Info. Theory, vol. IT-51, pp. 2867–2877, August 2005.
• [7] R. Zamir, S. Shamai (Shitz), and U. Erez, “Nested linear/lattice codes for structured multiterminal binning,” IEEE Trans. on Info. Theory, vol. IT-48, pp. 1250–1276, June 2002.
• [8] Y. Steinberg and S. Shamai (Shitz), “Achievable rates of the broadcast channel with states known at the transmitter,” in Proc. of IEEE Int. Symp. on Info. Theory, Adelaide, Australia, August 2005.
• [9] Y. Cemal and Y. Steinberg, “The multiple-access channel with partial state information at the encoders,” IEEE Trans. on Info. Theory, vol. IT-51, pp. 3992–4003, November 2005.
• [10] Y.-H. Kim, A. Sutivong, and S. Sigurjonsson, “Multiple user writing on dirty paper,” in Proc. IEEE Int. Symp. on Info. Theory, Chicago-USA, June 2004, p. 534.
• [11] A. Zaidi and L. Vandendorpe, “Rate regions for the partially-cooperative relay-broadcast channel with non-causal side information,” in Proc. of IEEE Int. Symp. on Info. Theory, Nice, France, June 2007, pp. 1246–1250.
• [12] S. Kotagiri and J. N. Laneman, “Achievable rates for multiple access channels with state information known at one encoder,” in Proc. Allerton Conf. Communications, Control, and Computing, Monticello, IL, 2004.
• [13] ——, “Multiple access channel with state information known at some encoders,” EURASIP Journal on Commnunication and Networking, 2008.
• [14] S. Kotagiri and J. Laneman, “Multiaccess channels with state information known to one encoder: A case of degraded message sets,” in Proc. IEEE Int. Symp. on Info. Theory, Nice, France, 2007.
• [15] A. Somekh-Baruch, S. Shamai (Shitz), and S. Verdù, “Cooperative encoding with asymmetric state information at the transmitters,” in Proc. IEEE Int. Symp. on Info. Theory, Nice, France, 2007.
• [16] ——, “Cooperative multiple access encoding with states available at one transmitter,” submitted to IEEE Trans. on Info. Theory, January 2007.
• [17] A. Zaidi and L. Vandendorpe, “Bounds on the capacity of the relay channel with states at the source,” in Proc. IEEE International Symposium on Information Theory and its Applications, ISITA (to appear), Auckland, New Zealand, Dec. 2008.
• [18] A. B. Carleial, “Multiple-access channels with different generalized feedback signals,” IEEE Trans. on Info. Theory, vol. IT-28, pp. 841–850, November 1982.
• [19] T. M. Cover and A. El Gamal, “Capacity theorems for the relay channel,” IEEE Trans. on Info. Theory, vol. IT-25, pp. 572–584, Sep. 1979.
• [20] A. Zaidi, S. Kotagiri, J. N. Laneman, and L. Vandendorpe, “Cooperative relaying with non-causal state available at the relay,” IEEE Trans. on Info. Theory (preprint), 2008.
• [21] R. G. Gallager, Information Theory and Reliable Communication.   New York: John Willey, 1968.
• [22] N. Merhav and S. Shamai (Shitz), “Cooperative encoding with asymmetric state information at the transmitters,” in Proc. IEEE Int. Symp. on Info. Theory, Seattle, USA, 2006, pp. 1184–1188.
You are adding the first comment!
How to quickly get a good reply:
• Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
• Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
• Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
The feedback must be of minimum 40 characters and the title a minimum of 5 characters