Optimal Correlators and Waveforms for Mismatched Detection
Abstract
We consider the classical Neymann–Pearson hypothesis testing problem of signal
detection, where under the null hypothesis (), the received signal is
white Gaussian noise, and under the alternative hypothesis (), the received
signal includes also an
additional non–Gaussian random
signal, which in turn
can be viewed as a deterministic waveform plus zero–mean, non-Gaussian
noise. However, instead of the classical
likelihood ratio test detector, which might be difficult to implement, in general,
we impose a (mismatched) correlation detector, which is relatively easy to implement, and
we characterize the optimal correlator weights in the sense of the best trade-off
between the false-alarm error exponent and the missed-detection error
exponent. Those optimal correlator weights depend (non-linearly, in general) on the underlying
deterministic waveform under . We then assume that the deterministic
waveform may also be free to be optimized (subject to a power constraint),
jointly with the correlator, and show that both the optimal waveform and the
optimal correlator weights may take on values in a small finite set of
typically no more than two to four levels, depending on the distribution of
the non-Gaussian noise component. Finally, we outline an extension of the
scope to a wider class of detectors that are based on linear combinations of
the correlation and the energy of the received signal.
Index terms: hypothesis testing, signal detection, correlation–detection, error exponent.
The Andrew & Erna Viterbi Faculty of Electrical and Computer Engineering
Technion - Israel Institute of Technology
Technion City, Haifa 32000, ISRAEL
E–mail: [email protected]
1 Introduction
The topic of detection of signals corrupted by noise has a very long history of active research efforts, as it has an extremely wide spectrum of engineering applications in the areas communications and signal processing. These include radar, sonar, light detection and ranging (LIDAR), object recognition in images and video streams, diagnosis based on biomedical signals, watermark detection in images and audio signals, seismological signal detection related to geophysical activity, and object detection using multispectral/hyperspectral imaging, just to name a few. One of the most problematic and frequently encountered issues in signal detection scenarios is mismatch between the signal model and the detector design, which is based upon certain assumptions on that model. Accordingly, the topic of mismatched signal detection has received considerable attention in the literature, see, e.g., [1], [10], [11], [18], [19], [20], [21], [28], and [30], for a non-exhaustive list of relevant references. The common theme in most of these works is the possible presence of uncertainties in the desired signal to be detected, in the steering vector, in the transfer function of the propagation medium, and/or in the distributions of the various kinds of noise, interference and clutter. Accordingly, adaptive detection mechanisms with tunable parameters have been developed and proposed in order to combat those types of mismatch.
Another line of earlier relevant research activity is associated with the notion of robust detection techniques, where the common theme is generally directed towards a worst-case design of the detector against small non-parametric uncertainties around some nominal noise distribution, most notably, a Gaussian distribution. See, e.g., [2], [4], [5], [9], [12], [13], [15], [16], [17], [22], [24], and [25]. See also [14] for a survey on the subject.
Last but not least, when the uncertainty is only in a finite number of parameters of the model, the problem is normally treated in the framework of composite hypothesis testing, where the popular approach is the well–known generalized likelihood ratio test (GLRT) [27], which is often (but not always) asymptotically optimal in the error exponent sense, see, for example, [3], [6], [7], and [29]. The GLRT is applied also in some of the above cited articles on mismatched detection, among many others. Another approach to composite hypothesis testing is the competitive minimax approach, proposed in [8].
Our objective in this work is partially related to those studies, but it is different. It is associated with mismatched detection, except that the origin of this mismatch is not quite due to uncertainty in the signal-plus-noise model, but it comes from practical considerations: the optimal likelihood ratio test (LRT) detector might be difficult to implement in many application examples, especially in the case of sensors that are built on small, mobile devices which are subjected to severe limitations on power and computational resources. In such situations, it is desirable that the detector would be as simple as possible, e.g., a correlation detector, or a detector that is based on correlation and energy. Within this framework, the number of arithmetic operations (especially the multiplications) should be made as small as possible. Clearly, a detector from this class cannot be optimal, unless the noise is Gaussian, hence the mismatch. Nonetheless, we would like to find the best correlator weights in the sense of optimizing the trade-off between the false-alarm (FA) and the missed–detection (MD) rates. This would partially compensate for the mismatch in case the noise is not purely Gaussian.
More precisely, consider the following signal detection problem, of distinguishing between two hypotheses:
(1) | |||
(2) |
where is an independently identically distributed (IID), zero-mean Gaussian noise process with variance , independent of , which is another random process, that we decompose as , with being a deterministic waveform and being an IID, zero-mean noise process, which is not necessarily Gaussian in general. The non–Gaussian noise component, , can be thought of as signal–induced noise (SIN), which may stem from several possible mechanisms, such as: echos of the desired signal, multiplicative noise, cross-talks from parallel channels conveying correlated signals, interference by jammers, and in the case of optical detection using avalanche photo-diodes (APDs), it corresponds to shot noise plus multiplicative noise due to the random gain of the device (see, e.g., [23] and references therein, for more details). In general, may also designate randomness that could be attributed to uncertainty associated with the transmitted signal.
As mentioned above, the optimal LRT detector might be considerably difficult to implement in practice since the probability density function (PDF) of under involves the convolution between the Gaussian PDF of and the (non-Gaussian) PDF of , which is typically complicated. As said, a reasonable practical compromise, valid when the underlying signal is not identically zero, is a correlation detector, which compares the correlation, , to a threshold, where are referred to as the correlator weights, and the threshold controls the trade-off between the FA probability and the MD probability. Our first objective is to characterize the best correlator weights, , in the sense of the optimal trade-off between the FA probability and the MD probability, or more precisely, the optimal trade-off between the asymptotic exponential rates of decay of these probabilities as functions of the sample size , i.e., the FA exponent and the MD exponent. Clearly, the optimal correlation detector is, in general, not as good as the optimal, LRT detector, but it is the best compromise between performance and the practical implementability in the framework of correlation detectors. As very similar study was already carried out in [23], in the context of optical signal detection using photo-detectors, where the optimal correlator waveform was characterized in terms of the optical transmitted signal in continuous time, and was found to be given by a certain non-linear function of the optical signal.
Here, we study the problem in a more general framework, in the sense that the PDF of the SIN, , is arbitrary. Moreover, we expand the scope in several directions, in addition to the study that is directly parallel to that of [23].
-
1.
We consider the possibility of limiting the number of levels to be finite (e.g., binary, ternary, etc.), with the motivation of significantly reducing the number of multiplications needed to calculate the correlation, .
-
2.
We jointly optimize both the signal and the correlator, . Interestingly, here both the optimal signal and the optimal correlation weights turn out to have a finite number of levels even if this number is not restricted a-priori. The number of levels depends on the PDF of , and it is typically very small (e.g., two to four levels). Moreover, the optimal and turn out to be proportional to each other, in contrast to the non-linear relation resulting when only is optimized while is given.
-
3.
We outline an extension to a wider class of detectors that are based on linear combinations of the correlation, , and the energy, , with the motivation that it is, in fact, the structure of the optimal detector when is Gaussian noise, and that it is reasonable regardless, since the under the power (or the variance) of the received signal is larger than under .111In fact, when , the correlation term becomes useless altogether and the energy term becomes necessary. We also address the possibility of replacing the energy term by the sum of absolute values, , which is another measure of signal intensity, with the practical advantage that its calculation does not require multiplications.
The outline of the remaining part of this work is as follows. In Section 2, we formalize the problem rigorously and spell out our basic assumptions. In Section 3, we characterize the optimal correlator, , for a given signal, , subject to the power constraint. In Section 4, we address the problem of joint optimization of both and , both under power constraints, and finally in Section 5, we outline extensions to wider classes of detectors that are based on correlation and energy.
2 Assumptions and Preliminaries
Consider the signal detection model described in the fifth paragraph of the Introduction. We assume that are independent copies of a zero–mean random variable (RV), , that has a symmetric222The symmetry assumption is imposed mostly for convenience, but the results can be extended to address also non-symmetric PDFs. PDF, , around the origin, and that it has a finite cumulant generating function (CGF),
(3) |
at least in a certain interval of the real valued variable . Note that since
is assumed symmetric
around the origin, then so is . We also assume
that is twice differentiable within the range it exists. It is well
known to be a convex function, because its second derivative cannot be
negative, as it can be viewed
as the variance of under the PDF that is proportional to .
Further assumptions on and its CGF will be
spelled out in the sequel, at the places they will be needed. The following simple special cases
will accompany our derivations and discussions repeatedly in the sequel:
Case 1. is zero-mean, Gaussian RV with variance :
(4) |
Case 2. is a Laplacian RV with parameter , i.e., :
(5) |
Case 3. in a binary RV, taking values in with equal probabilities:
(6) |
Case 4. is a uniformly distributed RV across the interval :
(7) |
The signal vector, , , , is assumed known, and we denote its power by , that is,
(8) |
Consider the class of correlation detectors, i.e., detectors that compare the correlation, , to a certain threshold , where is a vector of real valued correlator coefficients, henceforth referred to as the correlator, for short. The decision rule is as follows: If , accept the null hypothesis, , otherwise, accept the alternative, . The threshold, , controls the trade-off between the FA probability and the MD probability of the detector. To allow exponential decay (as grows without bound) of both types of error probabilities, we let vary linearly with , and denote , where is a real valued constant, independent of .
In order to have a well-defined asymptotic FA exponent, we assume that the correlator, , has a fixed power,
(9) |
which is independent of , or, more generally, that the right–hand side (RHS) of eq. (9) tends to a certain fixed positive power level, as (otherwise, the normalized logarithm of the FA probability would oscillate indefinitely, without a limit). Indeed, the FA probability of the correlation detector is given by
(10) |
where is the well-known -function,
(11) |
and denotes equivalence in the exponential scale, in other words, the notation, , for two positive sequences and , means that . It follows from (10) that the FA exponent is given by
(12) |
Thus, the FA exponent depends on only via . It follows that for a given , if we wish to achieve a given, prescribed FA exponent, (where is a given positive number), we must have
(13) |
In other words, a constraint on the FA exponent means a corresponding constraint on the asymptotic power of to be no larger than .
In order to have a well–defined MD exponent, our assumptions concerning the asymptotic behavior of and will have to be more restrictive: We will assume that as , the pairs obey a certain joint PDF, , in the following sense: For every ,
(14) | |||||
where denotes expectation with respect to (w.r.t.) . Whenever there is no room for confusion, the subscript will be omitted and the expectation will be denoted simply by . The function will be referred to as the asymptotic empirical joint PDF of and .333In the sequel, we will encounter one scenario where the asymptotic empirical PDF will be irrelevant, but this scenario will be handled separately, in the original domain of vectors of dimension .
The MD probability is now upper bounded, exponentially tightly, by the Chernoff bound, as follows. Denoting the Gaussian random variable, , we have
(15) | |||||
where the last step is due to the symmetry of the . The resulting MD exponent is therefore given by
(16) |
which is a functional of .
The problem of optimal correlator design for a given , is equivalent to the problem of finding a conditional density, , that maximizes the MD exponent subject to the power constraint, . The problem of joint design of both and is asymptotically equivalent to the problem of maximizing the MD exponent over subject to the power constraints, and and , for some given . The first problem is relevant when the detector designer has no control of the transmitted signal, for example, when the transmitter and the receiver are hostile parties, which is typically the case in military applications. The second problem is relevant when the transmitter and the receiver cooperate. In radar applications, for example, the transmitter and the receiver are the same party. In sections 3 and 4, we address the first problem and the second problem, respectively.
Comment 1. Instead of maximizing the MD exponent for a fixed threshold, , and a fixed power constraint, , in order to fit a prescribed FA exponent, in principle, there is an alternative approach: maximize the MD exponent directly for a given FA exponent, by substituting in the MD exponent expression. Not surprisingly, in this case, the MD exponent would become invariant to scaling in (as any scaling of can be absorbed in for all terms of the MD exponent), and so, there would be no need for the -constraint, but this invariance property holds only after maximizing over , not for a given . However, maximizing over as a first step of the calculation, does not seem to lend itself to closed form analysis, in general, and consequently, it would make the subsequent optimization extremely difficult, if not impossible, to carry out. We therefore opt to fix both and throughout our derivations.
3 Optimum Correlator for a Given Signal
In view of the discussion in Section 2, we wish to find the optimal conditional density, , in the sense of maximizing
(17) |
subject to the power constraint,
(18) |
At this stage, we carry out this optimization for a given , but with the understanding that eventually, will be subjected to optimization as well. To this end, let us denote the derivative of by , and for a given , define the function
(19) |
Observe that since is convex, is monotonically non-decreasing, and so, is monotonically strictly increasing, which in turn implies that it has an inverse. We denote the inverse of by . Also, since is assumed zero mean, then , and hence also and . Note also that (and hence also ) is a linear function if and only if is Gaussian. The following lemma characterizes the optimal .
Theorem 1
Let the assumptions of Section 2 hold. Assume further that is such that there exists (possibly, depending on ), with . Otherwise, if , set . Then, the optimal conditional density, , is given by
(20) |
where is the Dirac delta function.
The theorem tells that the best correlator, , for a given , is obtained by the relation,
(21) |
which means that is given by a function of , which is non-linear unless is Gaussian. To gain an initial insight regarding the condition on , consider the Gaussian example (Case 1, eq. (4)). In this case, , and so, , whose power is for , which is non-negative as long as . In general, the exact choice of is not crucial, as the prescribed FA exponent can be achieved by adjusting proportionally to . However, once is chosen, we will keep it fixed throughout (see Comment 1 above).
Proof of Theorem 1. Consider the following chain of inequalities and inequalities.
(22) | |||||
where (a) is since the objective is affine in both and in (and hence is concave in and
convex in ), (b) is since the unconstrained maximum of the conditional expectation
of a function of given is attained
when puts all its mass on the maximum of this function, (c)
is because the maximum is over a concave function of , which is attained at the point of zero-derivative,
, (d) is by the postulate that , and (e) is because either or
. The upper bound on the constrained maximum in the first line of the
above chain is therefore
attained by with probability one, which is equivalent to (20).
This completes the proof of Theorem 1.
Optimal correlator weights within a finite set. There is a practical motivation to consider the case where is restricted to be a binary vector with bipolar components, taking the values and only. The reason is that in such a case, the implementation of the correlation detector involves no multiplications at all, as it is equivalent to the comparison of the difference
to . Here, the maximization over (step (b) in the proof of Theorem 1) is carried out just over its two allowed values, and . As is symmetric, the maximum is readily seen to be attained by , which means .
Suppose, more generally, that is constrained to take on values in a finite set whose cardinality is fixed, independent of . This can be considered as a compromise between the above two extremes of performance and computational complexity, since the number of multiplications need not be larger than . The design of such a signal is very similar to the scalar quantizer design problem: A finite–alphabet signal is defined as follows. Let , where , , and let , , be given. Define
(23) |
for some given . We wish to minimize
(24) |
over , , and , . The necessary conditions for optimality are obtained by equating to zero all partial derivatives w.r.t. , , and , . This results in the following sets of equations:
Alternatively, we may represent these equations as:
(25) | |||||
(26) |
where is tuned such that
(27) |
as before. The first set of equations is parallel to the nearest-neighbor condition in optimal quantizer design, and the second set corresponds to the centroid condition. Optimal signal design can be conducted iteratively, like in quantizer design, by alternating between the two sets of equations.
Example 1. Consider the case where (i.e., Case 1). In this case, , and so, , which leads to
(28) |
and so,
(29) |
Choosing
(30) |
yields
(31) |
which results in
(32) |
If is constrained to be binary, then as we already saw, and then
(33) |
For the more general quantization, we obtain
(34) |
For simplicity, let us assume that is uniform across the interval . In this case, , and so,
(35) |
It follows that have uniform spacings across the support of , that is, , . Accordingly,
(36) |
where is chosen such that
(37) |
The binary case considered above is the special case of . This concludes Example 1.
If is itself a finite–alphabet signal, then the optimal is also a finite-alphabet signal with the same alphabet size, even without restricting it to be so in the first place. If this alphabet is small enough and/or there is a strong degree of symmetry, one might as well optimize the levels of directly subject to the power constraint. Consider, for example, the case of a 4-ASK signal, , for some given . Then, since the PDF of is assumed symmetric, the alphabet of the optimal must be of the form for some . Assuming that along half of the time and along the other half, then and in the corresponding halves, and so, , or . Thus, the MD exponent should be maximized over one parameter only (beyond the optimization over ), which is . In particular,
(38) | |||||
We next examine this expression in several examples.
Example 2. Let be a binary symmetric source, taking values for some (Case 3). Then, owing to eq. (6), eq. (38) becomes
(40) | |||||
The ‘classical’ correlator, where , corresponds to the choice instead of maximizing over . In Fig. 1, we compare the two curves of the MD exponent as functions of . Since they share the same level of , the FA exponents are the same for a given . As can be seen, the optimal correlator significantly outperforms the classical one, which is optimal in the Gaussian case only. This concludes Example 2.

Example 3. We conduct a similar comparison for the case where is distributed uniformly over (Case 4), which corresponds to eq. (7). The results are displayed in Fig. 2, and as can be seen, here too, the optimal correlator significantly improves upon the classical one. This concludes Example 3.

It is interesting to note that in both Examples 2 and 3, for large , the two graphs approach each other
faster than they approach zero. A possible intuitive explanation is that for large
, what counts is the behavior of the PDF of , fairly close to
its peak, where the regime of the central limit theorem is quite relevant, and so, there is no significant
difference from Case 1, where is Gaussian and the classical correlator
is good. Mathematically, as grows, the optimum decreases, and
so, it ‘samples’ the function in the vicinity of the origin, where it is
well approximated by a quadratic function, just like in the Gaussian case (Case 1).
Example 4. Finally, consider the case where is Laplacian (Case 2). In this case, the differences turned out to be rather minor – see Fig. 3. A plausible intuition is that the Laplacian PDF is much ‘closer’ to the Gaussian PDF, relative to the binary distribution and the uniform distribution of Examples 2 and 3. This concludes Example 4.

The loss relative to the optimal LRT detector depends on the relative intensity of the process compared to the Gaussian noise component.
4 Joint Optimization of the Correlator and the Signal
So far, we have concerned ourselves with the optimization of the correlator waveform, for a given signal, . But what would be the optimal signal (subject to a power constraint) when it is jointly optimized with ? Mathematically, we are interested in the problem,
(41) | |||||
where in (a) we have used the simple fact that, for a given and , the correlation, is maximized by . Earlier, we maximized the MD exponent w.r.t. for a given and found that the optimal is given by a function, , which is, in general, non–linear (unless is Gaussian). Now, on the other hand, the optimal for a given turns out to be given by a linear function. These two findings settle together if and only if takes values only in the set of solutions, , to the equation
(42) |
for some (and then takes the corresponding values according to their relationship). The two sides of the equation represent the non-linear and the linear relations, respectively. Note that always includes at least the solution . Once is chosen, is allowed to take on values only within . The inner minimization over in the last line of (41) is obviously lower bounded by , which is defined as
(43) |
where , and is understood to be a weight function over , i.e., for all and . While this expression appears complicated, there are two facts that help to simplify it significantly. The first is that, in most cases, is a finite set (unless is linear, or contains linear segments), and the second is that only two members of suffice, i.e., eq. (43) simplifies to
(44) |
The function has the flavor of a lower convex envelope for the function , but with the exception that the support of the convex combinations is limited to . Finally, the optimal MD exponent is given by
(45) |
The optimal is one that achieves for the maximizing and , that is, the components of take only two values in , with relative frequencies given by and , where and are the achievers of , being the optimal . In other words, the optimal signal has at most four levels, and , for some and .
Comment 2. By a simple change of variables, , it is readily seen that depends on and only via the quantity , and so, it might as well be denoted as .
Observe that while the function is always convex, nothing general can be asserted regarding convexity or concavity properties of the function , as the internal square root function, which is concave, may or may not destroy the convexity of the composite function, depending on the function . In other words, may either be convex, or concave, or neither. For example, if with equal probabilities (as in Case 3), then which is concave in . On the other hand, if is Laplacian with parameter (Case 2), then , which is convex in . By mixing these two distributions, we can also make it neither convex, nor concave, as will be shown in the sequel.
Let us examine now several special cases, where the form of can be determined more explicitly.
1. Consider first the Gaussian case (Case 1), where , namely, it is linear in . In this case, for , , the choice of is immaterial, and . In this case, any signal with power is equally good, as expected.
2. Consider next the case where is is convex. Then,
(46) | |||||
(47) | |||||
(48) |
where the inequality is achieved with equality whenever with probability one, and then this constant must be . So, here , and , which is expected, because in the convex case, there is no need for any non-trivial convex combinations. The optimal signal vector is any member of , and then is the corresponding member of . It is interesting to note that they both turn out to be DC or bipolar signals, which is good news from the practical point of view, as discussed in Section 1.
3. We now move on to the case where is concave. In this case, it is instructive to return temporarily to the original domain of vectors of finite dimension , find the optimal solution in that domain, and finally, take the limit of large (see footnote no. 3). We therefore wish to minimize s.t. . Since and each is limited to the range , we can lower bound the function (which is concave as a function of ), by a linear function of , as follows:
(49) |
with equality at and . Consequently,
(50) | |||||
(51) | |||||
(52) | |||||
(53) |
with equality if one of the components of is equal to and all other components vanish, and then, the same component of is (and, of course, all other vanish), correspondingly. Here, we have and . Asymptotically, as grows without bound, , and the limit exists since is monotonically non-increasing by the assumed concavity of . If this limit happens to vanish (like in Case 3, for instance), then the interference has no impact whatsoever on the MD exponent for the optimal and . Here too, the optimal signaling is binary.
We now summarize our findings, in this section so far, in the following theorem.
Theorem 2
Let the assumptions of Section 2 hold. Then, and are proportional to each other with and taking values in a finite set of size at most two (), and the MD exponent is given by eq. (45).
-
1.
If the function is convex, then both and are either DC or bipolar, and the MD exponent is given by
(54) -
2.
If the function is concave, then the components of both and are all zero, except for one component which exploits their entire energy. The MD exponent is given by
(55)
Finally, we should consider the case where is neither convex, nor concave. Here, we will not carry out the full calculations needed, but we will demonstrate that may include more than one positive solution, in addition to the trivial solution at the origin. Consider, for example a mixture of the binary PDF and the Laplacian PDF with weights and , respectively (). In this case,
(56) |
If is close to 1, the hyperbolic cosine term is dominant for small and moderate values of , where is concave. In contrast, when approaches , the second term tends steeply to infinity and hence must be convex. So in this example, is concave in a certain range of relatively small and at some point it becomes convex. Now, the derivative w.r.t. is given by
(57) |
As discussed above, the first step is to solve the equation
(58) |
As there is no apparent analytical closed-form solution to this equation, here we demonstrate the solutions graphically. In Fig. 4, we plot the functions and vs. for the following parameter values: , , , , and . As can be seen, in this example, there are two positive solutions (in addition to the trivial solution, ), which are approximately, and . Thus, in this case, , which corresponds to the set of power levels, . According to the above discussion, optimal signaling is associated with time-sharing between two out of these three signal levels. Given this simple fact, the optimal signal levels, say, and , and the optimal weight parameter, , can also be found directly by maximizing the MD error exponent expression with respect to these parameters, subject to the power constraint, , similarly as was done earlier in the example of the 4-ASK signal in Section 3.

5 Detectors Based on Linear Combinations of Correlation and Energy
In this section, we provide a brief outline of a possible extension of the scope to a broader class of detectors that compare the test statistic
to a threshold, . The motivation stems from the fact that the two hypotheses, and , differ not only in the presence of the signal, , but also in the presence of the SIN, , which adds to the energy (or the variance) of the received signal. In the extreme case, where , the simple correlation detector we examined so far (corresponding to ) would be useless, but still, one expects to be able to distinguish between the two hypotheses thanks to the different energies of the received signal. Indeed, if is Gaussian white noise (Case 1), the optimal LRT detector obeys this structure with .
For practical reasons, it would also be relevant to consider detectors that are based on
where the second term is another measure of the signal intensity, but with the advantage that its calculation does not require multiplications. We shall consider both classes of detectors, but provide merely the basic derivations of the MD exponent, without attempt to arrive at full, explicit solutions. Nevertheless, we will make an attempt to make some observations on the solutions.
We begin with the first class of detectors mentioned above. The FA probability is readily bounded by
(59) | |||||
which depends on only via , as before.
As for the MD probability, we define
(60) |
and
(61) |
Then,
(62) | |||||
where and (a) is due to the identity
(63) |
which is the characteristic function of a zero-mean Gaussian random variable with variance .444Alternatively, it can be viewed as the Fourier transform relation between two Gaussians, one in the domain of and one in the domain of . We now define
and we arrive at the following expression of the MD exponent:
(64) | |||||
where . Note that this expression is of the same form of the one we had earlier, except that is replaced by , is replaced by , and is replaced by .555It is easy to verify that is convex in simply because is such. Therefore, its derivative is monotonically non-decreasing. This expression should now be jointly maximized w.r.t. subject to the power constraints, , . Using the same techniques as before, it is not difficult to infer that the optimal for a given is linear in , whereas the optimal for a given is given by a non-linear equation. Whenever the number of simultaneous solutions to both equations is finite, the signal levels can be optimized directly, as before. As for the optimization of , among all pairs that give rise to the same value of the FA exponent, one chooses the one that maximizes the MD exponent.
Moving on to the second class of detectors, the analysis can be carried out using the same technique as above, where the this time, we use the Fourier transform identity,
(65) |
which, as before, enables to exploit the independence between and once the expectation operator is commuted with the inverse Fourier transform integral over . Equipped with this identity, we have
(66) | |||||
Thus, defining
(67) |
the MD exponent is
(68) |
However, in this case, there is an additional complication, which stems from the fact that the FA exponent depends on not only via . A standard Chernoff-bound analysis yields
(69) | |||||
Therefore, the maximization of the MD exponent will have to incorporate the full asymptotic PDF of and not just its second moment.
References
- [1] F. Bandiera, D. Orlando, and G. Ricci, Advanced Radar Detection Schemes Under Mismatched Signal Models, Synthesis Lectures in Signal Processing, Morgan & Claypool Publishers, 2009.
- [2] J. Capon, “On the asymptotic efficiency of locally optimum detectors,” IRE Trans. Inform. Theory, pp. 67–71, 1961.
- [3] E. Conte and G. Ricci, “Sensitivity study of GLRT detection in compound Gaussian clutter,” IEEE Trans. on Aerospace and Electronic Systems, vol. 34, no. 1, pp. 308–316, January 1998.
- [4] A. H. El-Sawy and V. D. Vandelinde, “Robust detection of known signals,” IEEE Transactions on Information Theory, vol. IT–23, no. 6, pp. 722–727, November 1977.
- [5] A. H. El-Sawy and V. D. Vandelinde, “Robust sequential detection of signals in noise,” IEEE Transactions on Information Theory, vol. IT–25, no. 3, pp. 346–353, November 1979.
- [6] E. Erez and M. Feder, “The generalized likelihood ratio decoder can be uniformly improved for Gaussian intersymbol interference channels,” Proc. ISITA 2000, Honolulu, Hawaii, November 2000.
- [7] E. Erez and M. Feder, “Uniformly improving the generalized likelihood ratio test for linear Gaussian channels,” Proc. Allerton Conference on Communications Control and Computing, pp. 498–507, September 2000.
- [8] M. Feder and N. Merhav, “Universal composite hypothesis testing: a competitive minimax approach,” IEEE Trans. Inform. Theory, special issue in memory of Aaron D. Wyner, vol. 48, no. 6, pp. 1504–1517, June 2002.
- [9] E. A. Geraniotis, “Performance bounds for discrimination problems with uncertain statistics,” IEEE Transactions on Information Theory, vol. IT–31, no. 5, pp. 703–707, September 1985.
- [10] F. Gini, M. V. Greco, A. Farina, and P. Lombardo, “Optimum and mismatched detection against -distributed plus Gaussian clutter,” IEEE Trans. Aerospace and Electronic Systems, vol. 34, no. 3, pp. 860–876, July 1998.
- [11] C. Hao, B. Liu, S. Yan, and L. Cai1, “Parametric adaptive Radar detector with enhanced mismatched signals rejection capabilities,” EURASIP Journal on Advances in Signal Processing, vol. 2010, Article ID 375136, 2010.
- [12] S. A. Kassam, “Robust hypothesis testing for bounded classes of probability densities,” IEEE Transactions on Information Theory, vol. IT–27, no. 2, pp. 242–247, March 1981.
- [13] S. A. Kassam, G. Moustakides, and J. G. Shin, “Robust detection of known signals in asymmetric noise,” IEEE Transactions on Information Theory, vol. IT–28, no. 1, pp. 84–91, January 1982.
- [14] S. A. Kassam and H. V. Poor, “Robust techniques for signal processing: a survey,” Proc. IEEE, vol. 73, no. 3, pp. 433–481, March 1985.
- [15] S. A. Kassam and J. B. Thomas, “Asymptotically robust detection of a known signal in contaminated non-Gaussian noise,” IEEE Transactions on Information Theory, vol. IT–22, no. 1, pp. 22–26, January 1976.
- [16] S. M. Kay, “Robust detection by autoregressive spectrum analysis,” IEEE Trans. on Acoustics, Speech, and Signal Processing, vol. ASSP–30, no. 2, pp. 256–269, April 1982.
- [17] V. M. Krasnenker, “Stable (robust) detection methods for signal against a noise background,” Automatation Remote Control, vol. 41, no. 5, pt. 1, pp. 640–659, May 1980.
- [18] J. Liu and J. Li, “Robust detection in MIMO Radar with steering vector mismatches,” IEEE Trans. Signal Processing, vol. 67, no. 20, October 2019.
- [19] W. Liu, J. Liu, Y. Gao, G. Wang, and Y. Wang, “Multichannel signal detection in interference and noise when signal mismatch happens,” Signal Processing, vol. 166, 107268 2020.
- [20] W. Liu, W. Xie, R. Li, F. Gao, X. Hu, and Y. Wang, “Adaptive detection in the presence of signal mismatch,” Journal of Systems Engineering and Electronics, vol. 26, no. 1, pp. 38–43, February 2015.
- [21] W. Liu, W. Xie, and Y. Wang, “Parametric detector in the situation of mismatched signals,” IET Radar, Sonar and Navigation, vol. 8, no. 1, pp. 48-–53, 2014.
- [22] R. D. Martin and S. C. Schwartz, “Robust detection of a known signal in nearly Gaussian noise,” IEEE Transactions on Information Theory, vol. IT–17, no. 1, pp. 50–56, January 1971.
- [23] N. Merhav, “Optimal correlators for detection and estimation in optical receivers,” IEEE Trans. Inform. Theory, vol. 67, no. 8, pp. 5200–5210, August 2021.
- [24] G. V. Moustakides, “Robust detection of signals: a large deviations approach,” IEEE Transactions on Information Theory, vol. IT–31, no. 6, pp. 822–825, November 1985.
- [25] G. V. Moustakides and J. B. Thomas, “Min-max detection of weak signals in phi-mixing noise,” IEEE Transactions on Information Theory, vol. IT–30, no. 3, pp. 529–537, May 1984.
- [26] X. Shuwen, S. Xingyu, and S. Penglang, “An adaptive detector with mismatched signals rejection in compound Gaussian clutter,” Journal of Radars, vol. 8, no. 3, pp. 326-334, 2019.
- [27] H. van Trees, Detection, Estimation and Modulation Theory, part I, John Wiley & Sons, New York 1968.
- [28] L. Wei-jian, W. Li-cai, D. Yuan-shui, J. Tao, X, Dang, and W. Yong-liang, “Adaptive energy detector and its application for mismatched signal detection,” Journal of Radars, vol. 4, no. 2, pp. 149–159, April 2015.
- [29] O. Zeitouni, J. Ziv, and N. Merhav, “When is the generalized likelihood ratio test optimal?” IEEE Trans. Inform. Theory, vol. 38, no. 5, pp. 1597–1602, September 1992.
- [30] D. Zhao, W. Gu, L. Jie, and L. Jing, “Weighted detector for distributed target detection in the presence of signal mismatch,” 2020 Proc. IET International Radar Conference (IET IRC 2020), pp. 960–963, doi: 10.1049/icp.2021.0816.