Coding for classicalquantum channels with rate limited side information at the encoder: An informationspectrum approach
Abstract
We study the hybrid classicalquantum version of the channel coding problem for the famous Gel’fandPinsker channel. In the classical setting for this channel the conditional distribution of the channel output given the channel input is a function of a random parameter called the channel state. We study this problem when a rate limited version of the channel state is available at the encoder for the classicalquantum Gel’fandPinsker channel. We establish the capacity region for this problem in the informationspectrum setting. The capacity region is quantified in terms of spectralsup classical mutual information rate and spectralinf quantum mutual information rate.
1 Introduction
In traditional information theory literature it is common to study the underlying problems assuming that the channel characteristics do not change over multiple use. The proofs appeal to typicality of sequences or typical subspaces in the quantum setting [1]: the empirical distribution of symbols in a long sequence of trials will with high probability be close to the true distribution [2]. However, information theoretic arguments based on typicality or the related Asymptotic Equipartition Property (AEP) assume that both the source and channel are stationary and/or ergodic (memoryless), assumptions that are not always valid, for example, in [3] Gray analyzes the details of asymptotically mean stationary sources, which are neither stationary nor ergodic. To overcome such assumptions Verdú and Han pioneered the technique of informationspectrum methods in their seminal work [4]. In this work Verdú and Han define the notions of limit inferior and limit superior in probability. They then use these definitions to establish the capacity of general channels (channels that are not necessarily stationary and/or memoryless). Since this work of Verdú and Han there have been a considerable interest in generalizing the results of information theory in the information spectrum setting, see for example, [5, 6, 7, 8] and references therein.
This general technique of informationspectrum methods wherein no assumptions are made on the channels and/or sources were extended to the quantum case by Hayashi, Nagaoka and Ogawa. Using this method they studied the problem of quantum hypothesis testing [9, 10], deriving the classical capacity formula of general quantum channels [11] and establishing general formula for the optimal rate of entanglement concentration [12]. Since the work of Hayashi, Nagaoka and Ogawa the study of various quantum information theoretic protocols in the information spectrum setting have been one of the most interesting areas of research in the theoretical quantum information science. In [13] Bowen and Datta further carried forward this approach to study various other quantum information theoretic problems. In [14] Datta and Renner showed that there is a close relationship with the information theoretic quantities that arise in the informationspectrum scenario and smooth Rényi entropies which play a crucial role in oneshot information theory. In [15] Radhakrishnan, Sen and Warsi proved oneshot version of the Marton inner bound for the classicalquantum broadcast channels. They then showed that their oneshot bounds yields the quantum informationspectrum genralization of the Marton inner bound in the asymptotic setting.
In this paper, we carry forward the subject of studying quantum information theoretic protocols in the informationspectrum setting. We study the problem of communication over the sequence (also called as the classicalquantum Gel’fandPinsker channel), where are the input and state alphabets and is a positive operator with trace one acting on the Hilbert space We establish the capacity region of this channel when rate limited version of the state sequence is available at the encoder. Figure 1 below illustrates this communication scheme.
The classical version of this problem was studied by Heggard and El Gamal (achievability) in [16] in the asymptotic iid setting. They proved the following:
Theorem 1.
Fix a discrete memoryless channel with state characterized by Let be such that
for some distribution Then, the rate pair is achievable.
Furthermore, in [16] Heggard and El Gamal argued that Theorem 1 implies the result of Gel’fand and Pinsker [17] who showed the following:
Theorem 2.
Fix a discrete memoryless channel with state characterized by The capacity of this channel when the state information is directly available noncausally at the encoder is
where
The above formula for the capacity is quite intuitive. If we set and in Theorem 2 then we rederive the famous Shannon’s channel capacity formula [18]. However, when , Theorem 2 implies that there is a loss in the maximum transmission rate per channel use at which Alice can communicate to Bob. This loss in the transmission rate is reflected by the term . Thus, can be thought of as the minimum number of bits Alice needs to send to Bob per channel use to help him get some information about the channel state sequence . Bob can then use this information about to recover the intended message.
Our Result
We establish the capacity region of the classicalquantum Gel’fandPinsker channel in the informationspectrum setting when rate limited version of the channel state is available at the encoder. In the informationspectrum setting the channel output need not be a tensor product state. Furthermore, the channel state is a sequence of arbitrarily distributed random variables. This extremely general setting is the hallmark of informationspectrum approach. We prove the following:
Theorem 3.
Let be a sequence of classicalquantum Gel’fandPinsker channels. The capacity region for this sequence of channels with rate limited version of the channel state available only at the encoder is the set of rate pairs satisfying the following:
The information theoretic quantities are calculated with respect to the sequence of states where for every ,
An immediate consequence of Theorem 3 is the following corollary:
Corollary 1.

(Hayashi and Nagaoka, [11]) The capacity of a sequence of classicalquantum channels is the following:

The capacity of a sequence of classicalquantum Gel’fandPinsker channels with channel state directly available at the encoder is the following:
where for every ,
2 Definition
Definition 1.
Let be a sequence of pair of random variables, where for every and take values over the set . The spectralsup mutual information rate between and is defined as follows:
(1) 
where is arbitrary and the probability above is calculated with respect to .
Definition 2.
Let and be sequences of quantum states where for every and are density matrices acting on the Hilbert space The spectralinf mutual information rate between and is defined as follows:
(2) 
where is arbitrary and represents a projection operator onto the nonnegative Eigen space of the operator
Definition 3.
An code for the Gel’fandPinsker channel with coded side information available at the encoder consists of

a state encoding

an encoding function (possibly randomized)

A decoding POVM such that
Definition 4.
A rate pair is achievable if there exists a sequence of codes such that
The set of all achievable rate pairs is known as the capacity region.
3 Proof of Theorem 3
3.1 Achievability
Let,
(3)  
(4)  
(5)  
(6) 
Let,
(7) 
where is calculated with respect to the sequence of states and Further, for every let
(8) 
Fix Define the following sets:
Furthermore, let and be defined as follows:
(9)  
(10) 
In what follows we will use the notation to represent the set
The codebook: Let be as in the statement of the theorem. Let , so that . Let be drawn independently according to the distribution . We associate these samples with a row vector having entries. We then partition this row vector into classes each containing elements. Every message is uniquely assigned a class. We will denote the class corresponding to the message by .
Fix Further, let be drawn independently according to the distribution We will denote this collection of sequences by These collection of sequences present in are made known to Alice as well.
Charlie’s encoding strategy: For each let be independently and uniformly distributed over For a given realisation of the state sequence , let be an indictor random variable defined as follows:
(11) 
Further, for a given realisation of the state sequence let be an indicator random variable defined as follows:
(12) 
where and are defined in (9) and (10). Charlie on observing the state sequence finds an index such that . If there are more than one such indices then is set as the smallest one among them. If there is none such index then Charlie then sends this index to Alice.
Alice’s encoding strategy: For each pair , let be independently and uniformly distributed over and let be defined as follows:
(13) 
where is defined in (3) and is defined in (8). Let be an indicator random variable such that
(14) 
Further, let be an indicator random variable defined as follows:
(15) 
To send a message and on receiving the index from Charlie, Alice finds an index such that If there are more than such indices then is set as the smallest one among them. If there is none such index then Alice then randomly generates and transmits it over the classicalquantum channel over channel uses. In the discussions below we will use the notation to highlight the dependence of on . A similar encoding technique was also used by Radhakrishnan, Sen and Warsi in [15].
Bobs’ decoding strategy: For each we have the operators as defined in (8). Bob will normalize these operators to obtain a POVM. The POVM element corresponding to will be
(16) 
Bob on receiving the channel output measures it using these operators. If the measurement outcome is then he outputs if Similar decoding POVM elements were also used by Wang and Renner in [19].
Probability of error analysis: Let a message be transmitted by Alice by using the protocol discussed above and suppose it is decoded as by Charlie. We will now show that the probability averaged over the random choice of codebook, the state sequence and is arbitrary close to zero. By the symmetry of the code construction it is enough to prove the claim for There are following sources of error:

Charlie on observing the state sequence does not find a suitable such that .

Alice on receiving the index from Charlie is not able to find a suitable such that such that

Charlie finds a suitable and Alice finds a suitable , but Bob’s measurement is not able to determine the index correctly.
Let and be the indices chosen by Charlie and Alice. Let us now upper bound the probability of error while decoding the transmitted message. Towards this we first define the following events:
We now have the following bound on the error probability:
(17) 
where the first inequality follows from the setting of the protocol discussed above and remaining all of the inequalities till (17) follow from the union bound. In what follows we will now show that for large enough we have
where is arbitrarily close to zero such that .
Consider
(18) 
where follows because are independent and identically distributed and are independent and identically distributed according to the distribution follows from the definition of follows from the inequality follows becuase follows because and union bound and follows because is large enough such that Let us now bound the each of the last two terms on the R.H.S. of (3.1) as follows:
Consider :
(19) 
where follows from Markov inequality; follows from the definition of and taking expectation over the random variable and follows under the assumption that is large enough such that
Consider :
(20) 
where follows from Markov inequality; follows from the definition of and by taking the expectation over the random variable and follows because of the following set of inequalities:
where follows from the Markov inequality, follows from the definition of mentioned in (8), follows from the definition of mentioned in (4) and follows under the assumption that is large enough such that Thus, it now follows from (3.1), (19) and (20) that
(21) 
Consider