Capacity of Gaussian
Arbitrarily-Varying Fading Channels
Abstract
This paper considers an arbitrarily-varying fading channel consisting of one transmitter, one receiver and an arbitrarily varying adversary. The channel is assumed to have additive Gaussian noise and fast fading of the gain from the legitimate user to the receiver. We study four variants of the problem depending on whether the transmitter and/or adversary have access to the fading gains; we assume the receiver always knows the fading gains. In two variants the adversary does not have access to the gains, we show that the capacity corresponds to the capacity of a standard point-to-point fading channel with increased noise variance. The capacity of the other two cases, in which the adversary has knowledge of the channel gains, are determined by the worst-case noise variance as a function of the channel gain subject to the jammer’s power constraint; if the jammer has enough power, then it can imitate the legitimate user’s channel, causing the capacity to drop to zero. We also show that having the channel gains causally or non-causally at the encoder and/or the adversary does not change the capacity, except for the case where all parties know the channel gains. In this case, if the transmitter knows the gains non-causally, while the adversary knows the gains causally, then it is possible for the legitimate users to keep a secret from the adversary. We show that in this case the capacity is always positive.
Index Terms: Gaussian arbitrarily-varying fading channel, Gaussian arbitrarily-varying channel, fast fading channel, capacity, active adversary
I Introduction
It is a matter of great importance to study wireless communication channels since they include numerous challenges caused by noise and fading. The wireless environment also allows any uninvited signal to enter the channel. These signals can act as interference or in the worse case as a malicious jammer whose aim is to disrupt the communication between the legitimate users. In this work, we explore how these various signals interact with each other to restrict the overall capacity of the channel.
Goldsmith and Varaiya in [2] studied the point-to-point Gaussian channel with fast fading in which the fading coefficients are drawn i.i.d. from a distribution known to all parties. They determined the capacity of the channel where the fading coefficients are available at the receiver and possibly the transmitter. When the fading coefficients are not available at the transmitter, the capacity is equal to the expected value of the capacity of the corresponding Gaussian channel with the received signal-to-noise ratio affected by the fading gains. If both the transmitter and receiver know the exact channel gains, the transmitter can maximize the capacity by constructing its signal power as a function of the channel gains.
Another line of work studies channels in the presence of a malicious adversary. The adversary can be an active attacker who sends its signal to the channel in order to cease or restrict the communication between the legitimate users. If the adversary’s signal is arbitrarily chosen or is drawn from an unknown distribution, then the channel is called as an arbitrary-varying channel (AVC). We focus on the variant of the problem wherein the adversary’s knowledge is limited to the code of the legitimate user, but it does not have access to the user’s transmitted messages, neither exact nor noisy. Csiszár and Narayan established the capacity of discrete AVC with input and jammer power constraints in [3, 4]. They also derived the capacity for a continuous version of the AVC with input and state (jammer) power constraints in [5].
Csiszár and Narayan in [5] characterized the capacity of the Gaussian arbitrarily-varying channel under the average probability of error and the average power constraints for input and state. It is assumed that the adversary does not have any information about the legitimate signal except the code. It is shown that if the adversary has enough power to forge a message and send it to the channel, then the receiver gets confused and cannot distinguish between the true message and the malicious one. This occurrence is called symmetrizability , causing the capacity to drop to zero. In [5], it is shown that the adversary can symmetrize the channel if and only if it has greater power that the legitimate transmitter. However, if the jammer does not have enough power, then the capacity is equal to the capacity of a standard Gaussian channel with the noise variance increased by the power of the jammer.
The problem of AVC capacity has been also studied for discrete and continuous multi-user channels. The authors in [6] considered the discrete arbitrarily-varying multiple-access memoryless channel. We characterized lower and upper bounds for the capacity of arbitrarily-varying Gaussian interference channel in [7]. List-decoding is also investigated for the discrete and Gaussian AVCs in [8, 9] and in [10], respectively. The list capacity is derived by using list-decoding which decodes a list of messages instead of unique message at the receiver.
The three elements of Gaussian noise, fading, and adversary have been previously combined to study problems with a (passive) eavesdropping adversary, rather than an (active) transmitting adversary. A secrecy capacity problem with slow fading, in which the fading gains are constant over each block length, is considered in [11]. The authors determined the secrecy capacity of the channel where both of transmitter and receiver know the channel state information (CSI) of the main path, but do not have any information about the eavesdropper channel. Furthermore, the capacity is generalized to multiple eavesdroppers in [12]. The problem is also studied in [13] for a specific case of a fast Rayleigh fading eavesdropper channel and a standard Gaussian main channel where the CSI are only known to the eavesdropper.
In this paper, we consider a Gaussian AVC with fast fading on the main path, as illustrated in Fig. 1; we refer to this channel as the Gaussian arbitrarily-varying fading channel (GAVFC). We characterize the capacity of the GAVFC under the average probability of error criterion. Similar to the Gaussian fading channel, we also assume that everyone knows the fading gain distribution including the adversary, but they may or may not know the realization of the gain sequence. Note that the “arbitrarily-varying” aspect of the channel is the adversary’s signal, not the channel gains, which we assume to be random from a known distribution. The receiver always needs the exact fading gains to decode the message, while the adversary and the transmitter may or may not know the exact values of fading gains. Therefore, we derive the capacity of the GAVFC for four cases wherein the channel gains are available at the transmitter and/or adversary as follows:
-
•
Neither the transmitter nor the adversary knows the channel gains.
-
•
Only the transmitter knows the channel gains.
-
•
Only the adversary knows the channel gains.
-
•
Both the transmitter and the adversary knows the channel gains.
If the jammer does not know the channel gains, we show that the capacity is equal to the capacity of the corresponding fading channel with increased noise variance by the power of the jammer. If the jammer knows the fading gains, then it can choose its signal as a function of the gains, and under some power constraints it can symmetrize the channel and make the capacity zero. Note that if the channel gains are not available at the adversary, it does not have the required channel information to symmetrize the channel. Moreover, all the results still hold if the adversary and the encoder have the channel gains causally or non-causally, except one situation. If the adversary knows the channel gains causally while the encoder knows them non-causally, then the adversary cannot symmetrize the channel since the encoder possesses some extra information that the adversary does not.
The rest of the paper is organized as follows. We describe the GAVFC model and define the various capacities in Sec. II. We state our main theorem including the capacity of the GAVFC in all cases in Sec. III. Before giving the proof of our main theorem, we need some auxiliary lemmas and tools which are presented in Sec. IV. Later, in sections V, VI, VII, and VIII, we provide the converse and achievability proofs of each of the main results in Theorem 1. Finally in the Appendix, we provide a brief proof for the auxiliary results.
Notation: We use bold letters to indicate -length vectors. We employ and to denote inner product and Hadamard product (element-wise multiplication), respectively. We indicate the positive-part function, 2-norm and the expectation by , and , respectively. Also, for an integer , stands for the set . Notation represents the identity matrix of size . Each of and functions has base 2. Moreover, , and denotes Gaussian random variable with mean and variance .
II Problem Statement

The Gaussian arbitrarily-varying fading channel (GAVFC) in Fig. 1 is a point-to-point fading channel with additive Gaussian noise and an intelligent adversary who does not have any information about the transmitted signal except the code. The received signal is given by
(1) |
where is a random sequence of identical and independently distributed (i.i.d.) fast fading channel gains from the legitimate transmitter to the receiver drawn from continuous distribution assumed to have positive and finite variance, is the -length deterministic vector representing the user’s signal, is the adversary signal chosen arbitrarily, and is a random -length noise vector distributed as a sequence of i.i.d. zero mean Gaussian random variables with variance , independent of , and . Note that the receiver always knows the exact fading coefficients while the transmitter and the adversary may not know the gains, know them causally, or know them non-causally.
Define an code for the GAVFC by a message set, an encoding function and a decoding function as follows:
-
•
Message set ,
-
•
Encoding function (one of the following)
-
–
(No knowledge) where ,
-
–
(Causal) where and for ,
-
–
(Non-causal) where and for ,
-
–
-
•
Decoding function ,
where the rate of the code is . The message is drawn uniformly from the set . If the encoder does not know the channel gains, it maps the message to . If the encoder knows the channel gains causally, then it maps the message to , and if the encoder knows the channel gains non-causally, then it maps the message to where . Given channel gains at the receiver, the signal is decoded by function to the message . Moreover, we assume that if the channel gains are available at the transmitter then the transmitter’s signal satisfies the expected power constraints for any message . Otherwise, the power constraint is . The same definition applies to the adversary’s signal power constraint, i.e. if the adversary knows the channel gains, the constraint is ; otherwise, it is . The three parameters , , and as well as the distribution of fading gains are known to all parties.
The probability of error for the message in the presence of adversary signal is now given by the probability that . Thus, the average probability of error for a specific is
(2) |
If the adversary knows the channel gains non-causally then his signal is denoted by for . Alternatively, if the adversary knows the gains causally, then the adversary’s action is given by functions for where and . Therefore, the average probability of error for this specific is
(3) |
Finally, the overall probability of error is maximized over all possible choices of jammers’ sequences which satisfy either or . Rate is achievable if there exists a sequence of codes where . The capacity is the supremum of all achievable rates. We denote the capacity of the GAVFC as where denotes the transmitter’s knowledge, and denotes the adversary’s knowledge; and can be U, C, or N depending on whether the transmitter or adversary does not know the gains (U unknown), knows the gains causally (C), or knows the gains non-causally (N). For example, is the capacity where the transmitter does not know the gains and the adversary knows the gains non-causally.
III Main Results
We present our results for the capacity of GAVFC whether the fading channel gains are available causally or non-causally at the encoder and/or the adversary (the decoder always knows the gains) in the following theorems.
Theorem 1
The capacities of the GAVFC are given by
(4) | |||
(5) | |||
(6) | |||
(7) | |||
(8) |
Note that when the encoder knows the gains (5), (7), and (8), the capacity expression includes a maximization of the input power as a function of the gain, similar to the result in [2]. Similarly, when the jammer knows the gains (6)–(8), the capacity expression includes a minimization that represents the jammer’s choice of noise power as a function of the gain. Moreover, when the jammer knows the gains, with enough power it can symmetrize the channel by mimicking the legitimate signal, thus reducing the capacity to zero. However, in (8) we have assumed that the adversary knows the gains causally and the encoder and the decoder know the gains non-causally. Thus, the encoder and decoder effectively share a secret (the channel gains at the end of the block) unknown to the adversary, so the adversary cannot symmetrize the channel. It is also worth mentioning that for the other cases (except (8)) our proof works exactly the same whether the transmitter and/or the adversary know the gain sequence causally, non-causally, or even memorylessly (i.e., at time , you only know the gain value at time ).
While we have stated the theorem by writing the capacities in terms of optimization over the and/or functions, these expressions can be computed by solving for the optimizing functions. In particular, the optimum value of in (5) is where is obtained by , and the optimum value of in (6) is a function of gain as follows
(9) |
and is obtained by solving . Moreover, the optimum values of and in (7) are
(10) | ||||
(11) |
where , and are found by solving , and , respectively. Finally, the optimum values of and in (8) are
(12) | ||||
(13) |
where and can be obtained by solving and , respectively.
In the Fig. 2, the capacity of GAVFC with Rayleigh fading is shown for and the Rayleigh distribution scale parameter whether the channel gains are available at the encoder and/or the adversary. However, is the capacity of Gaussian arbitrary-varying channel with no fading. Note that when the transmitter’s knowledge increases, the capacity increases, whereas when the adversary’s knowledge increases, the capacity decreases. On the other hand, the knowledge of adversary about the channel gains may decrease the capacity, and in this case if the adversary’s power exceeds , the capacity will be zero by the symmetrizability.
The proofs of the different capacity variants all follow a similar pattern, so we have attempted to reduce the redundancy in our presentation. We provide essentially one converse proof for each capacity expression in the theorem:
-
1.
Sec. V-A: Converse for .
-
2.
Sec. VI-A: Converse for . This also bounds which is trivially upper bounded by .
-
3.
Sec. VII-A: Converse for . This also bounds which is trivially upper bounded by .
-
4.
Sec. VIII-A: Converse for . This also bounds which is trivially upper bounded by . Essentially the same proof also works for . The case requires a different proof also covered in this section.
We provide essentially three achievability proofs:
-
1.
Sec. VI-B: Achievability for . This also bounds the , which is trivially lower bounded by . It also provides a bound for , because the same proof works assuming (i.e., the encoder’s power is independent of the channel gain).
-
2.
Sec. VIII-B: Achievability for . This also bounds and , which are trivially lower bounded by . It also provides a bound for and , because the same proof works again assuming .
-
3.
Sec. VIII-C: Achievability for . This case is different from all others in that there is effectively a shared secret between encoder and decoder.

IV Auxiliary Results and Tools
Before proceeding to the proofs, we first define the typical set for continuous random variables with probability density function as follows:
(14) |
where is the differential entropy of . Next, we define the typical set for continuous random variables with probability density function and a discrete random variable with probability mass function as follows:
(15) |
where and denote the entropy of and the conditional differential entropy of given .
Throughout the achievability proofs, we will utilize several lemmas including the joint typicality lemma and conditional typicality lemma for Gaussian random variables given in [10]. In addition, we will need the following two lemmas; they show that with high probability a Gaussian codebook satisfies several desirable properties. The proof is given in the Appendix.
Lemma 2
Fix . There exists such that the following holds. Let for , be a zero mean Gaussian codebook with variance . Let be drawn from probability density function . With probability approaching 1 as , for any where , there exists a function such that
(16) |
where the union is over zero mean conditionally Gaussian random vectors given .
Lemma 3
Fix . There exists such that the following holds. Let for , be a zero mean Gaussian codebook with variance . Let be drawn from probability density function . With probability approaching 1 as , for any
-
•
zero-mean conditionally Gaussian random vector given where and ,
-
•
where ,
there exists a function such that
(17) | |||
(18) | |||
(19) | |||
(20) |
V Capacity Proof with Gains Available at Decoder
V-A Converse Proof
We initially assume that for any arbitrary adversary strategy there is a sequence of codes with vanishing probability of error. The adversary can generate a Gaussian sequence with variance for any ; if this sequence has power less than , it is transmitted, otherwise, the adversary sends the all-zero sequence. Note that the power of this Gaussian sequence exceeds only with small probability by the law of large numbers. With this choice of adversary, the channel corresponds to a standard Gaussian fading channel with the noise variance where the channel gains are available only at the decoder. Therefore, using capacity of a non-adversarial Gaussian fading channel [14] for arbitrarily small , we may upper bound the capacity by
(21) |
V-B Achievability Proof
The achievability proof of this case can be counted as a special case of in Sec. VI-B where both encoder and decoder know the channel gains. However, in this case since the encoder does not know the channel gains, we do not have any function at the encoder. In other words, the achievability proof for this case is identical to that in Sec. VI-B with .
VI Capacity Proof with Gains Available at Encoder and Decoder
VI-A Converse Proof
As in the previous case, the adversary can simply send Gaussian send noise with variance . By the law of large numbers, the resulting channel is equivalent to a standard Gaussian fading channel with the knowledge of gains at both encoder and decoder and noise variance with high probability. Thus, since can be chosen arbitrarily small, from the capacity of a non-adversarial Gaussian fading channel [2], we have
(22) |
VI-B Achievability Proof
For simplicity we assume . Suppose any arbitrary function that satisfies and . We further assume that has a positive variance. Note that this is only a concern if the optimum ; in this case, we can instead take where are two positive constants and can be chosen arbitrarily small. Let
(23) |
We now propose a code sequence, and prove that using this code the probability of error tends to zero as .
Codebook generation: Fix . We generate i.i.d zero mean Gaussian sequences with variance for each . By Lemma 2 and Lemma 3, we assume that the deterministic codebook satisfies (16)–(20).
Encoding: Since the transmitter knows the channel gains, it sends (at time signal is sent) if its power is less than , otherwise it sends zero.
Decoding: Given , let be the set of messages such that for some random variables , and zero mean Gaussian where are mutually independent.
Now, we define the decoding function as
(24) |
Analysis of the probability of error: Suppose the true message sent by the legitimate user is message with the power constraint . Then, the overall probability of error is upper bounded by where
(25) | ||||
(26) |
Consider any state sequence . By (16), with high probability where are independent, and . By the conditional typicality lemma, for every with high probability where are mutually independent, and . Thus, according to the definition of , with high probability and tends to zero as .
Define the shorthand . Let be a finite -dense subset in the set of all distributions of random vectors that are determined by and jointly zero mean Gaussian vector independent of with bounded covariances at most . Note that because the distribution of is fixed, the overall distribution of can be determined by the covariance matrix of , so only needs to cover a compact set. Now, we may upper bound by
(27) |
where
(28) |
We will show that for all vectors and all vectors which are Gaussian given (whether or not they are in ). Let . We may restrict ourselves to where
(29) | |||
(30) | |||
(31) | |||
(32) | |||
(33) |
where (29) holds since the input , adversary , fading gains and noise are all generated independently, (30)–(31) follows from , and , (32) holds since we have are mutually independent using , and (33) corresponds to which is less than from (28).
Observe that if , then we would have
(34) | ||||
(35) | ||||
(36) | ||||
(37) |
where (34) follows from (32), (36) holds because are all mutually independent by the assumption and (29), and the last equality holds since is independent of and because . Therefore, .
Moreover, from (28) we have
(38) | ||||
(39) | ||||
(40) | ||||
(41) |
where (40) holds because , are mutually independent, are mutually independent, and are independent by (29), (30) and the assumption . Canceling from both sides of (41) gives us
(42) |
Now, if we apply the result from (37) to (42), we get
(43) | ||||
(44) | ||||
(45) | ||||
(46) |
which is a contradiction since we assume is always positive. Thus, there exists an such that
(47) |
Also, by (20), we may restrict ourselves to distributions where
(48) |
and
(49) |
Note that . We also have the upper bound
(50) | ||||
(51) |
where (51) follows from , (19) and the joint typicality lemma.
Now, let us consider three cases as follows:
Case (a): that implies . From (51), for any
(52) | ||||
(53) | ||||
(54) | ||||
(55) | ||||
(56) |
where (56) follows from (47), (48) and (49). Therefore, vanishes exponentially fast if is sufficiently small.
Case (b): . Since and , from (49) we have
(57) | ||||
(58) | ||||
(59) |
Using this result in (48), we have
(60) |
Therefore,
(61) | ||||
(62) |
Now, from (51), we have for any
(63) | ||||
(64) | ||||
(65) |
where (64) follows from (62). We now lower bound as follows:
(66) | ||||
(67) | ||||
(68) | ||||
(69) | ||||
(70) | ||||
(71) | ||||
(72) |
where (72) follows from (32) and (33). Replacing this result in (65), we obtain
(73) |
meaning that is exponentially vanishing if is sufficiently small, and (23) holds.
VII Capacity Proof with Gains Available at Decoder and Jammer
VII-A Converse Proof
Consider a sequence of codes with vanishing probability of error that must function for arbitrary jamming signals. Because we are proving the converse, we may assume the best case scenario from the legitimate user’s perspective; in particular, that the adversary only knows the channel gains causally.
We begin with the case that . Given any function satisfying , we may obtain an upper bound by assuming that the jammer transmits a random sequence where is Gaussian with mean zero and variance for . Note that
(74) | ||||
(75) | ||||
(76) | ||||
(77) |
The resulting channel is equivalent to a standard Gaussian fading channel with the knowledge of gains only at the decoder and noise variance . From the capacity of a non-adversarial Gaussian fading channel
(78) |
Therefore, the capacity is also less than the minimum over all that satisfies .
(79) |
For the case , we first show that the adversary has enough power to choose a codeword and send it to the channel, thereby symmetrizing the channel. Let be a uniformly chosen message by the adversary and be the true message sent by the legitimate transmitter. Suppose the adversary chooses then the adversary’s power constraint is satisfied as follows:
(80) | ||||
(81) | ||||
(82) | ||||
(83) |
where (82) follows from the assumption , and (83) follows from the codebook power constraint . Given this choice of , . Thus, with high probability the decoder cannot decode the message since it does not know whether the true message is or .
VII-B Achievability Proof
The achievability proof of this case is very similar to the achievability proof of Sec. VIII-B where the encoder, the decoder and the adversary all know the channel gains. Here, the transmitter does not know the channel gain so it cannot leverage this knowledge to choose its transmit power. However, the achievability proof for this case is identical to that in Sec. VIII-B except that the transmitter’s power function is constant; i.e., .
VIII Capacity Proof with Gains Available at Encoder, Decoder, and Jammer
In this section, we first provide the converse proof for the case that the channel gains are available at the encoder, the decoder and the adversary in Sec. VIII-A. The converse proof includes all the four cases in which each of the adversary and the encoder knows the fading gains causally or non-causally. In Sec. VIII-B, we show the achievability proof of the case that the channel gains are available non-causally at the adversary and causally at the encoder. This proof also works for the two cases of channel gains being available causally at both the adversary and the encoder or non-causally at both ends. Finally, we provide the achievability proof for the last case when the channel gains are causally available at the adversary and non-causally available at the encoder in Sec. VIII-C.
VIII-A Converse Proof
Consider a sequence of codes with vanishing probability of error. Since in this case both the encoder and the adversary know the channel gains, we consider four cases to prove the converse whether each of them knows the fading gains causally or non-causally.
First assume that both the encoder and adversary know the channel gains causally. Let and where , for . Thus, satisfies as follows:
(84) | ||||
(85) | ||||
(86) |
where (86) follows by the power constraint for the input signal.
Now, similar to the previous case, where the adversary and decoder know the channel gains, we also have symmetrizability and non-symmetrizability cases, but with different conditions. We first show the symmetrizability case, that is if , then the jammer can symmetrize the channel. Suppose the adversary chooses a message uniformly at random and sends where for . Note that this selection of jamming signal is a causal function of the channel gains. Then we have
(87) | ||||
(88) | ||||
(89) | ||||
(90) | ||||
(91) | ||||
(92) | ||||
(93) |
Therefore, this choice of jammer satisfies the adversary power constraint. Given , the decoder cannot determine the correct message between true message or the adversary message with high probability. Thus, the probability of error is bounded away from zero. By the above argument, if for all where , then the capacity cannot be positive; the adversary can always symmetrize the channel, so the capacity is 0.
On the other hand, consider the case where there exists some function where and . Let be given by
(94) |
Since the transmitted codes should work for arbitrary jamming signals, an outer bound may be obtained by assuming the adversary sends . By the assumption that , the jammer’s expected power constraint is satisfied. Therefore, the rate is upper bounded by
(95) | ||||
(96) | ||||
(97) | ||||
(98) | ||||
(99) | ||||
(100) |
where (97) follows since the mutual information is less than the capacity of equivalent standard fading channel with noise variance , and the gains being available at both encoder and decoder, (98) follows by the definition of , (99) follows by the concavity of with respect to and Jensen’s inequality, and (100) follows since we have established that satisfies and .
Moreover, if the encoder knows the channel gains causally, and the adversary knows them non-causally, then the adversary is stronger than in the previous case, so exactly the same bound holds. If both encoder and adversary know the channel gains non-causally, then we instead assume
(101) |
where and , so we get the same upper bound.
However, the case wherein the encoder knows the channel gains non-causally, and the adversary knows them causally is somewhat different. In this case, the encoder may send while the adversary does not have any access to to construct . Thus, it cannot do better than sending Gaussian noise. In this case, we derive only a converse bound based on the adversary sending Gaussian noise. Hence, we obtain the following bound:
(102) |
Note that here we are not making the assumption that .
VIII-B Achievability Proof (Gains Available Non-causally at Adversary and Causally at Encoder)
We first quantize in the following way. Fix . Given the assumption that has finite variance, there exists a real-valued random variable with a finite support such that is a deterministic function of and for each . We further assume that is the expected value of within each quantization set; that is, .
Without loss of generality, assume . Let be a rate and be any function satisfying
(103) | |||
(104) | |||
(105) |
We construct a code as follows:
Codebook generation: Fix . Generate i.i.d. zero mean Gaussian sequences with variance for each . By Lemmas 3 and Lemma 2, we may assume that the deterministic codebook satisfies (16)–(20).
Encoding: Given message and gain sequence , the transmitter computes from the quantization function, and then sends (at time signal is sent) if ; otherwise, it sends zero. Note that here we assume that the encoder knows the channel gains causally.
Decoding: Given and , let and be the set of messages such that where is the quantized random variable from and are some random variables that are conditionally Gaussian given with zero mean and covariance
(108) |
where . Note that the following can be shown from (108).
(109) | |||
(110) | |||
(111) | |||
(112) |
Now, we define the decoding function as
(113) |
Analysis of the probability of error: Assume the legitimate transmitter sends message . Then, we can upper bound the probability of error by the summation of the following error probabilities:
(114) | ||||
(115) |
We can prove with high probability that
(116) | ||||
(117) | ||||
(118) | ||||
(119) |
where (117) follows from the law of large numbers for non-identical independent random variables , (118) follows from the assumption , and (119) follows from the assumption and .
Consider any jammer sequence . We may assume sequence is typical since it is drawn i.i.d. from the distribution . Similarly, is also typical because it is from the corresponding discrete distribution . Thus, is also typical with respect to some distribution where is conditionally Gaussian. Note that we can make no assumptions about the conditional variances defining , because the adversary is assumed to know in its choice of . By (16), with high probability where is independent of , and . Thus, by the conditional typicality lemma, with high probability where are independent of , and . Hence, using 119, we have where is sufficiently small compared to . Also, since , by (119) we may roughly assume and obtain . Moreover, to completely show that with high probability , we also need to compute the covariance matrix of given , where , and show that it is in the form of (108). First, since is independent of ,
(120) | ||||
(121) |
where follows from the weak union rule since is independent of . Furthermore,
(122) | ||||
(123) | ||||
(124) | ||||
(125) |
where (124) follows from the weak union rule for independent of and independent of . Therefore, the conditional covariance matrix of can be obtain from , (121) and (125), and is the same as (108). Now, since and the conditional covariance matrix of satisfies (108), with high probability , and vanishes as .
Using (119) and triangle inequality, we may upper bound by the following:
(126) |
Define the shorthand . Let denote a finite -dense subset in the set of all distributions of random vectors that are determined by and a random vector distributed conditionally zero mean Gaussian given with bounded covariances at most . Note that because the distribution of is completely known, the overall distribution of can be determined by the conditional covariance matrix of given for each of the finitely many realizations, so only needs to cover a compact set. Now, we may upper bound by
(127) |
where
(128) |
We will show that for all vectors and all vectors which are Gaussian given (whether or not they are in ). Let . We may restrict ourselves to where
(129) | |||
(130) | |||
(131) | |||
(132) | |||
(133) | |||
(134) | |||
(135) | |||
(136) | |||
(137) |
Note that using (20), we only need to consider the distributions that satisfies (129). in addition, (130)–(131) are obtained by the definition of , (132) holds since the codebook , Gaussian noise and fading gains are generated independently, and the adversary signal may depend on but not the others, (133) follows from (109), (134) follows from the power constraints of the codebook, the adversary and the distribution of noise, (135)-(136) follows from (111)-(112), and (137) follows from (128). Let . Therefore, using (136) we have , and by (137) we get .
Observe that if , then we would have
(138) | ||||
(139) | ||||
(140) | ||||
(141) |
where (138) follows from (135), (140) follows from the assumption in which is independent of , and (141) holds since is independent of . Therefore, and the covariance matrix of given is equal to
(142) |
The determinant of is that should be non-negative since the covariance matrix must be positive semi-definite. Thus, its expectation is also non-negative:
(143) |
However, since , (143) contradicts the initial assumption on in (104). Thus, there exists such that
(144) |
where we have used the fact that .
Probability may be upper bounded by
(145) | ||||
(146) |
where (146) follows from (19) and the joint typicality lemma.
We consider the following two cases.
Case (a): . Applying this condition to (129), we get
(147) | ||||
(148) |
Since then . Considering (146), for any we have
(149) | ||||
(150) | ||||
(151) |
where (151) follows from (144) and (148). Therefore, vanishes exponentially fast if is sufficiently small.
Case (b): . Then we may apply this condition to (129) as
(152) | ||||
(153) | ||||
(154) |
Since , we may upper bound (146) by
(155) | ||||
(156) | ||||
(157) |
where by (133), we have . In the following, we find a lower bound for the mutual information in (157).
(158) | ||||
(159) | ||||
(160) | ||||
(161) | ||||
(162) | ||||
(163) |
where (158) follows from data processing inequality, (162) follows from standard argument for the capacity of Gaussian channel, and (163) follows from the definition of . Therefore, by the assumptions about and in (104)–(105), , so by (157) is exponentially vanishing if and are sufficiently small.
It is worth mentioning that this achievability proof also works for the case where both the adversary and encoder know the channel gains causally, or both know the gains non-causally. Since in all three cases the knowledge of the encoder is not more than the knowledge of the adversary, the jammer is able to impersonate the legitimate transmitter, and thereby symmetrize the channel, depending on the power allocation.
VIII-C Achievability Proof (Gains Available Causally at Adversary and Non-causally at Encoder)
In this case, both the encoder and the decoder know the channel gains non-causally meaning that they know the whole string including . However, the adversary only knows the gains causally, so at time it only has access to . Therefore, both the encoder and the decoder have some extra common information that the adversary does not know. In particular, the encoder and the decoder immediately know which the adversary knows only at time . Hence, we can leverage this common knowledge between the encoder and the decoder as common randomness that is unknown to the jammer. Moreover, by the assumption that is a continuous random variable with positive variance, in fact just has infinite entropy, and thus can be considered a source of an infinite number of bits of common randomness. Therefore, we proceed to provide an achievability proof where the encoder and decoder are assumed to share an infinite source of common randomness. However, note that implementing this approach would require measuring to an arbitrarily level of precision, which is not practical. Even so, the random code reduction technique of, for example, [15, Lemma 12.8], can be used to show that only bits of common randomness need to be extracted from (or perhaps for some ) in order to achieve the same rate.
A large amount of the achievability proof is identical to the Sec. VIII-B. The main difference is that the codebook is based on common randomness between encoder and decoder, so we denote the codebook as random variables which are independent from the jammer signal. As a consequence, the symmetrizability condition is not needed, and the proof is somewhat simpler. In particular, we fix a satisfying (103), and a rate satisfying (105), and we prove achievability as follows.
Codebook generation: Let be a Gaussian codebook with variance satisfying (16). This random codebook is generated from the infinite source of common randomness, so it is unknown to the adversary.
Encoding: Given message and gain sequence , the transmitter first computes from the quantization function, and then sends (at time signal is sent) if ; otherwise, it sends zero.
Decoding: Given and , let and let be the set of messages such that where is the quantized random variable from and are conditionally Gaussian given with zero mean and covariance matrix as follows:
(166) |
where . Note that the following can be shown from (166):
(167) | |||
(168) | |||
(169) | |||
(170) |
Now, we define the decoding function as
(171) |
Analysis of the probability of error: Assume the legitimate transmitter sends message . Then, we can upper bound the probability of error by the summation of the following error probabilities:
(172) | ||||
(173) |
Using the same argument in Sec. VIII-B, we can prove with high probability that tends to zero as . Using (119) and triangle inequality, we may upper bound by the following:
(174) |
Defining and the same as Sec. VIII-B, we may now upper bound by
(175) |
where
(176) |
and satisfies the same properties as in (129)–(137). Now, it suffices to show that vanishes for all typical vectors and all vectors which are Gaussian given (whether or not they are in ).
Using the joint typicality lemma in [14, Remark 2.2] we may upper bound in (128) (with the codewords and ) as follows:
(177) | ||||
(178) | ||||
(179) |
where in (178) we have used the fact that is independent of , and (179) follows from (167). We now lower bound the mutual information in (179) by the following.
(180) | ||||
(181) | ||||
(182) | ||||
(183) | ||||
(184) | ||||
(185) |
where (180) follows from data processing inequality, (184) follows from standard argument for the capacity of Gaussian channel, and (185) follows from the definition of . Therefore, by the assumptions about and in (103) and (105), , so by (179) is exponentially vanishing if and are sufficiently small.
IX Conclusion
This paper studied two phenomena together which are usually studied separately: namely, active adversaries and fading. We derived the capacity of Gaussian arbitrarily-varying fading channels where the adversary knows the transmitter’s code but not the exact transmitted signal. The adversary affects the capacity by increasing the noise variance by an amount related the adversary’s power. The capacity also depends on whether the transmitter and/or the adversary know the fading gains or not. The transmitter uses its knowledge to maximize the channel capacity while the adversary tries to minimize the capacity by its knowledge of the channel gains. Furthermore, if the adversary’s knowledge is at least that of the transmitter’s knowledge, then the adversary is able to make the capacity zero with enough power. In this paper, we have focused on the scenario where fading applies to the transmitter-to-receiver path, but not the adversary-to-receiver path. Future work could including considering fading along both paths. Such an alternatively model would present somewhat different challenges. Alternative directions include considering fading and adversaries in network settings, or an adversary with some direct control of the fading gains.
Acknowledgment
This material is based upon work supported by the National Science Foundation under Grant No. CCF-1453718.
Appendix A Proof of Lemma 3
In order to prove (16), we use our proof in [16, Lemma 6] for one codebook. Moreover, to obtain (19)–(20), we apply the corresponding proof of the equations in [8, Lemma 1] for Gaussian distributions. Note that [8] focuses on discrete alphabets, but the same proofs can be extended to Gaussian distributions by quantization of the set of continuous random variables in the following way.
Let be Gaussian i.i.d. -length random vectors (codebook) independent from each other with . First let be a typical realization of i.i.d. continuous random variable with probability density function . Next, we quantize the set of all , into a -dense subset . For a fixed , fix and a covariance matrix such that is a -dense subset of for such that , and is a -dense subset of for positive definite covariance matrices with diagonals at most .
Using the similar proof in [8, Lemma 1], we obtain for given and covariance matrix that the complement of each event in (19)–(20) happens with decreasingly doubly exponential probability for sufficiently large meaning that
(186) | |||
(187) |
Then, in order to complete the proof, since for any fixed the cardinality of finite set is only increasingly exponentially in , and the set is finite along with the doubly decreasing exponential probabilities in (186)–(187), we derive that with probability approaching to , all inequalities in (19)–(20) hold simultaneously for sufficiently large . Since these inequalities hold for every element in the finite sets and , then for any vector and any given covariance matrix (with ) which is not in its corresponding -dense subset, there exists a point in the corresponding -dense subset that is close enough to it (in its distance neighborhood). Now, by using the continuity properties of all sets, we may conclude that (19)–(20) hold also for any point which is not in the -dense subset.
References
- [1] F. Hosseinigoki and O. Kosut, “Capacity of gaussian arbitrarily-varying fading channels,” in 2019 53rd Annual Conference on Information Sciences and Systems (CISS), March 2019, pp. 1–6.
- [2] A. J. Goldsmith and P. P. Varaiya, “Capacity of fading channels with channel side information,” IEEE Transactions on Information Theory, vol. 43, no. 6, pp. 1986–1992, Nov 1997.
- [3] I. Csiszar and P. Narayan, “The capacity of the arbitrarily varying channel revisited: positivity, constraints,” IEEE Transactions on Information Theory, vol. 34, no. 2, pp. 181–193, March 1988.
- [4] ——, “Arbitrarily varying channels with constrained inputs and states,” IEEE Transactions on Information Theory, vol. 34, no. 1, pp. 27–34, Jan 1988.
- [5] ——, “Capacity of the Gaussian arbitrarily varying channel,” Information Theory, IEEE Transactions on, vol. 37, no. 1, pp. 18–26, Jan 1991.
- [6] J. A. Gubner, “On the capacity region of the discrete additive multiple-access arbitrarily varying channel,” IEEE Transactions on Information Theory, vol. 38, no. 4, pp. 1344–1347, July 1992.
- [7] F. Hosseinigoki and O. Kosut, “The gaussian interference channel in the presence of a malicious jammer,” in 2016 54th Annual Allerton Conference on Communication, Control, and Computing (Allerton), Sept 2016, pp. 679–686.
- [8] B. L. Hughes, “The smallest list for the arbitrarily varying channel,” IEEE Transactions on Information Theory, vol. 43, no. 3, pp. 803–815, May 1997.
- [9] A. D. Sarwate and M. Gastpar, “List-decoding for the arbitrarily varying channel under state constraints,” IEEE Transactions on Information Theory, vol. 58, no. 3, pp. 1372–1384, March 2012.
- [10] F. Hosseinigoki and O. Kosut, “Capacity of the gaussian arbitrarily-varying channel with list decoding,” in 2018 IEEE International Symposium on Information Theory (ISIT), June 2018, pp. 471–475.
- [11] J. Barros and M. R. D. Rodrigues, “Secrecy capacity of wireless channels,” in 2006 IEEE International Symposium on Information Theory, July 2006, pp. 356–360.
- [12] P. Wang, G. Yu, and Z. Zhang, “On the secrecy capacity of fading wireless channel with multiple eavesdroppers,” in 2007 IEEE International Symposium on Information Theory, June 2007, pp. 1301–1305.
- [13] Z. Li, R. Yates, and W. Trappe, “Achieving secret communication for fast rayleigh fading channels,” IEEE Transactions on Wireless Communications, vol. 9, no. 9, pp. 2792–2799, September 2010.
- [14] A. E. Gamal and Y. H. Kim, Network information theory. Cambridge University Press, 2011.
- [15] I. Csiszár and J. Körner, Information Theory: Coding Theorems for Discrete Memoryless Systems. Cambri, 2011.
- [16] F. Hosseinigoki and O. Kosut, “The Gaussian interference channel in the presence of malicious jammers,” [Online] arXiv:1712.04133, Dec. 2017, submitted to IEEE Trans. on Information Theory.