This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A Neural Network-Prepended GLRT Framework for Signal Detection Under Nonlinear Distortions

Rajeev Sahay,  Swaroop Appadwedula,  David J. Love,  and Christopher G. Brinton R. Sahay, D. J. Love, and C.G. Brinton are with the Elmore Family School of Electrical and Computer Engineering, Purdue University, West Lafayette, IN, 47907 USA. E-mail: {sahayr,djlove,cgb}@purdue.edu.S. Appadwedula is with MIT Lincoln Laboratory, Advanced SATCOM Group, Lexington, MA, 02421 USA. E-mail: [email protected] work was supported in part by MIT Lincoln Laboratory.DISTRIBUTION STATEMENT A. Approved for public release. Distribution is unlimited. This material is based upon work supported by the Department of the Navy under Air Force Contract No. FA8702-15-D-0001. Any opinions, findings, conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the Department of the Navy.
Abstract

Many communications and sensing applications hinge on the detection of a signal in a noisy, interference-heavy environment. Signal processing theory yields techniques such as the generalized likelihood ratio test (GLRT) to perform detection when the received samples correspond to a linear observation model. Numerous practical applications exist, however, where the received signal has passed through a nonlinearity, causing significant performance degradation of the GLRT. In this work, we propose prepending the GLRT detector with a neural network classifier capable of identifying the particular nonlinear time samples in a received signal. We show that pre-processing received nonlinear signals using our trained classifier to eliminate excessively nonlinear samples (i) improves the detection performance of the GLRT on nonlinear signals and (ii) retains the theoretical guarantees provided by the GLRT on linear observation models for accurate signal detection.

Index Terms:
Generalized likelihood ratio test, dense neural network, nonlinear signal processing, wireless communications.

I Introduction

Signal detection is a classical problem with several applications in wireless communications [1]. The general detection problem entails identifying the presence of a particular signal of interest (SOI) in a received transmission containing different sources of background noise and interference. Traditionally, with adequate assumptions about the distributions of the SOI and interference, statistical hypothesis testing has been used to derive robust and highly accurate detectors [2, 3].

Despite their robustness under classical linear operations, nonlinear transformations and clipping of a received signal significantly degrades the performance of linear test statistics derived from statistical hypothesis testing. In particular, and of special interest to our study, high-powered throughput signals that are outside the linear operating regime of an amplifier cause nonlinear signal distortions, which result in clipped signals leading to nonlinearities in the received passband signal before it is observed at complex baseband. Such high-powered signals that induce nonlinearity arise from sources of interference such as loud signals in audio applications and high signal-to-noise ratio (SNR) interferers that dominate the SOI in situations such as the near-far problem. Achieving accurate detection in such scenarios despite the nonlinearity caused by the system’s hardware is vital. Therefore, robust detection models, with performance guarantees similar to those provided by likelihood methods, are needed for signal detection in the presence of hardware-induced nonlinearities.

In this work, we consider editing a received wireless signal that has undergone a nonlinear transformation prior to applying the generalized likelihood ratio test (GLRT) to perform detection. We find that, under nonlinear distortions of a received signal, each sample has been altered to a different extent. For example, certain nonlinear samples are significantly distorted from their linear counterparts while other nonlinear samples remain close to their linear counterparts. Motivated by this, we train a lightweight (i.e., low-parameter) feedforward dense neural network (DNN) capable of classifying samples that have undergone excessive nonlinearity. At test time, we pre-process our received nonlinear signal by eliminating the time samples that are predicted to be highly distorted by the DNN. The GLRT is then applied to the pre-processed signal, with the highly nonlinear time samples eliminated, to perform detection. Our results verify that this method allows a linear test statistic to operate in nonlinear environments while also retaining the theoretical guarantees of the GLRT.

Related Work: Historically, test statistics derived from the GLRT, as well as similar statistical hypothesis testing methods [4], have been thoroughly studied and shown to be robust signal detectors in linear operations [5, 6] with Gaussian noise. In addition, the GLRT has also been shown to be effective when the distribution of the background noise in the received signals is non-Gaussian [7, 8]. Yet, the GLRT is prone to significant performance degradation when the received signal has undergone a nonlinear transformation, regardless of the noise distribution [9].

Contrary to the long-standing work on detection using likelihood methods, machine learning has recently been proposed for signal detection in wireless communications and cognitive radio environments [10, 11, 12]. In particular, these studies have demonstrated the versatility that deep learning is able to provide in dynamic environments over linear methods when analytically characterizing the received signal may not be possible. However, contrary to likelihood-based detection, deep learning does not provide theoretical guarantees for detection, making the interpretation of the detection result challenging. In this work, we employ neural networks in conjunction with a likelihood-based detection test to leverage its strong classification performance while retaining the theoretical guarantees provided by the GLRT.

Summary of Contributions: The main contributions of this work are as follows:

  • Nonlinear Sample Classifier: We propose prepending a GLRT detector with a neural network classifier to identify time samples that have undergone a nonlinear transformation (Sec. II).

  • Linear GLRT Operation in Nonlinear Environment: We show the effectiveness of our nonlinear sample classifier, in conjunction with the GLRT-derived linear test statistic, in performing robust detection in nonlinear environments (Sec. III).

II Methodology

II-A Signal Modeling

Refer to caption
Figure 1: Our detection framework for nonlinear signals consisting of a DNN prepending the GLRT. Our framework is designed for the detection of a SOI, a2a_{2}, when the received signals have undergone nonlinearity due to hardware constraints.

We consider two time-series signals, a1a_{1} and a2a_{2}, transmitted from two different directions, impinging an MM element antenna array. Throughout this work, we will denote the interference signal as a1a_{1} and the SOI, which we are interested in detecting, as a2a_{2}. A snapshot of the output at the antenna array at time tt is given by

𝐳(t)=𝐱(u1)a1(t)+𝐱(u2)a2(t)+𝐧(t),\displaystyle\mathbf{z}(t)=\mathbf{x}(u_{1})a_{1}(t)+\mathbf{x}(u_{2})a_{2}(t)+\mathbf{n}(t),\quad t=1,,L,\displaystyle t=1,\ldots,L, (1)

where 𝐳(t)M×1\mathbf{z}(t)\in\mathbb{C}^{M\times 1} is a column vector that denotes the time sample of the MM sensor outputs at time tt, 𝐱(u1)\mathbf{x}(u_{1}) and 𝐱(u2)\mathbf{x}(u_{2}) are the array response vectors of signals a1a_{1} and a2a_{2} arriving from directions u1u_{1} and u2u_{2} (assuming that the direction parameters u1u_{1} and u2u_{2} are scalars), respectively, and a1(t)a_{1}(t) and a2(t)a_{2}(t) are the amplitudes of a1a_{1} and a2a_{2}, respectively, at time tt. Lastly, 𝐧(t)𝒞𝒩(𝟎,σ2𝐈)\mathbf{n}(t)\sim\mathcal{C}\mathcal{N}(\mathbf{0},\sigma^{2}\mathbf{I}) is complex additive white Gaussian noise (AWGN) with variance σ2\sigma^{2}, and LL is total number of time samples.

Let 𝐙=[𝐳(1),,𝐳(L)]M×L\mathbf{Z}=[\mathbf{z}(1),\cdots,\mathbf{z}(L)]\in\mathbb{C}^{M\times L} denote the complete observed signal at the antenna array. Given 𝐙\mathbf{Z}, we are interested in determining when the SOI turns on (i.e., when a2a_{2} becomes active). Here, we define the null hypothesis, 0\mathcal{H}_{0}, as the case where a2a_{2} is absent in 𝐙\mathbf{Z} (i.e., a2=𝟎a_{2}=\mathbf{0}), and the alternate hypothesis, 1\mathcal{H}_{1}, as the case where a2a_{2} is present in 𝐙\mathbf{Z} (i.e., a2𝟎a_{2}\neq\mathbf{0}). Formally, we characterize the distribution of 𝐙\mathbf{Z} under each hypothesis as

𝐙{𝒞𝒩(𝟎,𝐑),0𝒞𝒩(𝐗𝐀𝐓,𝐑),1,\mathbf{Z}\sim\begin{cases}\mathcal{C}\mathcal{N}(\mathbf{0},\mathbf{R}),&\mathcal{H}_{0}\\ \mathcal{C}\mathcal{N}(\mathbf{XAT},\mathbf{R}),&\mathcal{H}_{1}\end{cases}, (2)

where 𝐑\mathbf{R} is the true, unknown covariance of 𝐙\mathbf{Z}. Now, under 1\mathcal{H}_{1}, we are exclusively considering a system with one SOI in 𝐙\mathbf{Z}. Here, 𝐗=[𝐱(u2)]M×1\mathbf{X}=[\mathbf{x}(u_{2})]\in\mathbb{C}^{M\times 1} is the antenna array response of the SOI, 𝐀1×S\mathbf{A}\in\mathbb{C}^{1\times S} is the row vector of signal amplitudes, and 𝐓S×L\mathbf{T}\in\mathbb{C}^{S\times L} is the matrix [𝟎𝐈S][\mathbf{0}\quad\mathbf{I}_{S}] (S=LT0S=L-T_{0}) corresponding to the time samples during which the SOI is active (i.e., when a2a_{2} is present in 𝐙\mathbf{Z}), thus yielding 𝐀𝐓=[0,0,,0,a2(T0),a2(T0+1),,a2(L)]\mathbf{AT}=[0,0,\ldots,0,a_{2}(T_{0}),a_{2}(T_{0}+1),\ldots,a_{2}(L)], where T0T_{0} denotes the time sample at which a2a_{2} becomes active in 𝐙\mathbf{Z}. Our objective is to arrive at a metric that will allow us to determine the active hypothesis (i.e., where the SOI became active in 𝐙\mathbf{Z}).

II-B GLRT-based Linear Detection

Given 𝐙\mathbf{Z}, we will determine the active hypothesis by utilizing the GLRT and optimizing over the empirical covariances. Denoting h0h_{0} and h1h_{1} as the complex Gaussian densities corresponding to the distributions under 0\mathcal{H}_{0} and 1\mathcal{H}_{1}, the GLRT, by definition, is

max𝐑,𝐀h1(𝐙;𝐑,𝐀)max𝐑h0(𝐙;𝐑)=(a)|𝐑^0|min𝐀|𝐑^1|\frac{\underset{\mathbf{R},\mathbf{A}}{\text{max}}\hskip 2.84526pth_{1}(\mathbf{Z};\mathbf{R},\mathbf{A})}{\underset{\mathbf{R}}{\text{max}}\hskip 2.84526pth_{0}(\mathbf{Z};\mathbf{R})}\stackrel{{\scriptstyle\mathclap{\mbox{(a)}}}}{{=}}\frac{|\hat{\mathbf{R}}_{0}|}{\underset{\mathbf{A}}{\text{min}}\hskip 2.84526pt|\hat{\mathbf{R}}_{1}|}
=|𝐙𝐙H|min𝐀|(𝐙𝐗𝐀𝐓)(𝐙𝐗𝐀𝐓)H|γ,=\frac{|\mathbf{ZZ}^{\text{H}}|}{\underset{\mathbf{A}}{\text{min}}\hskip 2.84526pt|(\mathbf{Z}-\mathbf{XAT})(\mathbf{Z}-\mathbf{XAT})^{\text{H}}|}\geq\gamma, (3)

where γ\gamma is the threshold parameter, H denotes the Hermitian transpose, tr()\text{tr}(\cdot) denotes the trace of a matrix, and |||\cdot| denotes the determinant. In (a), the maximization over the numerator and denominator covariances yield the corresponding empirical covariances 𝐑^0=𝐙𝐙H\hat{\mathbf{R}}_{0}=\mathbf{ZZ}^{\text{H}} and 𝐑^1=(𝐙𝐗𝐀𝐓)(𝐙𝐗𝐀𝐓)H\hat{\mathbf{R}}_{1}=(\mathbf{Z}-\mathbf{XAT})(\mathbf{Z}-\mathbf{XAT})^{\text{H}}. The quantity in (3) is equivalent to

|𝐗H(𝐙old𝐙oldH)1𝐗||𝐗H(𝐙𝐙H)1𝐗|γ.\frac{|\mathbf{X}^{\text{H}}(\mathbf{Z}_{\text{old}}\mathbf{Z}_{\text{old}}^{\text{H}})^{-1}\mathbf{X}|}{|\mathbf{X}^{\text{H}}(\mathbf{Z}\mathbf{Z}^{\text{H}})^{-1}\mathbf{X}|}\geq\gamma. (4)

We refer the reader to [13] for additional details on this GLRT derivation. In (4), we have introduced 𝐙oldM×k\mathbf{Z}_{\text{old}}\in\mathbb{C}^{M\times k}, which corresponds to a partition of subsequent time samples in 𝐙\mathbf{Z}. This definition assumes that the GLRT will be applied to partitions of 𝐙\mathbf{Z}. Specifically, 𝐙old\mathbf{Z}_{\text{old}} will be used in conjunction with the next partition of time samples, 𝐙newM×k\mathbf{Z}_{\text{new}}\in\mathbb{C}^{M\times k}, to determine if a new signal (the SOI) became active in 𝐙new\mathbf{Z}_{\text{new}} which was absent in 𝐙old\mathbf{Z}_{\text{old}} [6]. Formally, 𝐙old\mathbf{Z}_{\text{old}} and 𝐙new\mathbf{Z}_{\text{new}} are given by

𝐙=[𝐙old𝐳(1),,𝐳(k),𝐙new𝐳(k+1),,𝐳(2k),,𝐳(L)],\mathbf{Z}=\begin{bmatrix}\makebox[0.0pt][l]{$\smash{\underbrace{\phantom{\begin{matrix}\mathbf{z}(1),\cdots,\mathbf{z}(k),\end{matrix}}}_{\text{$\mathbf{Z}_{\text{old}}$}}}$}\mathbf{z}(1),\cdots,\mathbf{z}(k),\makebox[0.0pt][l]{$\smash{\underbrace{\phantom{\begin{matrix}\mathbf{z}(k+1),\cdots,\mathbf{z}(2k),\end{matrix}}}_{\text{$\mathbf{Z}_{\text{new}}$}}}$}\mathbf{z}(k+1),\cdots,\mathbf{z}(2k),\cdots,\mathbf{z}(L)\\ \end{bmatrix},

where 𝐙old\mathbf{Z}_{\text{old}} and 𝐙new\mathbf{Z}_{\text{new}} are shifted one time sample per iteration, and the GLRT is reevaluated, for all time samples.

In our case, where 𝐗\mathbf{X} is completely unknown, we can base detection on the equivalent monotonically related suboptimal statistic of the GLRT [6] given by

tr(𝐑^old1𝐑^new)γ,\text{tr}({\hat{\mathbf{R}}_{\text{old}}}^{-1}\hat{\mathbf{R}}_{\text{new}})\geq\gamma, (5)

where we will reject 0\mathcal{H}_{0} if γγ0\gamma\geq\gamma_{0} for some threshold γ0\gamma_{0}, which is empirically determined. Here, 𝐑^old=𝐙old𝐙oldH\hat{\mathbf{R}}_{\text{old}}=\mathbf{Z}_{\text{old}}\mathbf{Z}_{\text{old}}^{\text{H}} is the empirical covariance of 𝐙old\mathbf{Z}_{\text{old}}, and 𝐑^new=𝐙new𝐙newH\hat{\mathbf{R}}_{\text{new}}=\mathbf{Z}_{\text{new}}\mathbf{Z}_{\text{new}}^{\text{H}} is the empirical covariance of 𝐙new\mathbf{Z}_{\text{new}}. Intuitively, the GLRT solution scans the spatial (covariance) domain on blocks of time samples and results in a contrast peak (further shown and discussed in Sec. III) when a signal (i.e., the SOI) impinging the antenna array from a different direction than the interference turns on in 𝐑^new\hat{\mathbf{R}}_{\text{new}} that was not present in 𝐑^old\hat{\mathbf{R}}_{\text{old}}.

II-C GLRT Breakdown in Nonlinearity

Now, let us consider the effect of the GLRT test statistic given in (5) when a nonlinearity is induced on the received signal. In this case, we will assume that we receive f(𝐙):M×LM×Lf(\mathbf{Z}):\mathbb{C}^{M\times L}\rightarrow\mathbb{C}^{M\times L} instead of 𝐙\mathbf{Z}, where f()f(\cdot) is some unknown nonlinear function that is individually applied to each element in 𝐙\mathbf{Z}. Specifically, the nonlinearity effect is given by

f(𝐙)=f([𝐳(1),,𝐳(L)])=[f(𝐳(1)),,f(𝐳(L))].f(\mathbf{Z})=f([\mathbf{z}(1),\ldots,\mathbf{z}(L)])=[f(\mathbf{z}(1)),\ldots,f(\mathbf{z}(L))]. (6)

The element-wise effect of f()f(\cdot) is critical because it will affect each sample differently. For example, certain received time samples in 𝐙\mathbf{Z} will remain in the linear operating region and thus will not be significantly distorted by f()f(\cdot), whereas other samples will experience high degrees of clipping and distortion due to the nonlinearity. The clipping distortion of each antenna element can be observed in the eigenvalue spread of the covariance matrix, 𝐑^old\hat{\mathbf{R}}_{\text{old}}.

The induced nonlinearity directly changes the distribution of 𝐙\mathbf{Z} and, as a result, the initial assumptions about the distributions of the interference and SOI in (2) no longer hold. Furthermore, since the nonlinear function, f()f(\cdot), is unknown, we are unable to characterize the distribution of the signal after it has been transformed via f()f(\cdot), thus preventing us from deriving a new GLRT-based test statistic to operate on f(𝐙)f(\mathbf{Z}). The inability to analytically characterize nonlinearly-processed received signals motivates a data driven approach to signal detection (as described in Sec. II-D), where data samples can be augmented such that the GLRT is able to operate accurately.

II-D DNN-Based Nonlinear Sample Classification

To combat the breakdown of the GLRT solution on nonlinear signals, we implement a dense (fully connected) neural network (DNN) classification model, which employs the time samples comprising f(𝐙)f(\mathbf{Z}) for training. Specifically, our objective is to train the DNN to classify excessively nonlinear time samples, i.e., f(𝐳(t))f(\mathbf{z}(t)), using both a linear received signal, 𝐙\mathbf{Z}, and its corresponding nonlinear transformation, f(𝐙)f(\mathbf{Z}). At test time, when we only receive f(𝐙)f(\mathbf{Z}) and are unaware of 𝐙\mathbf{Z}, we will use the DNN to identify and eliminate excessively nonlinear samples from f(𝐙)f(\mathbf{Z}) before applying the GLRT test statistic for detection. Our overall DNN-prepended GLRT proposed framework is shown in Fig. 1.

For compatibility with real-valued DNNs, we will represent each time sample (column vector) as a real-valued vector consisting of the sample’s corresponding real and imaginary components. Specifically, we denote f(𝐳(t))f(\mathbf{z}(t)) as f(𝐳(t))2M×1f(\mathbf{z}(t))\in\mathbb{R}^{2M\times 1} with the form

f(𝐳(t))=[{f(𝐳1(t))}{f(𝐳M(t))}{f(𝐳1(t))}{f(𝐳M(t))}],f(\mathbf{z}(t))=\begin{bmatrix}\Re\{f(\mathbf{z}_{1}(t))\}\\ \vdots\\ \Re\{f(\mathbf{z}_{M}(t))\}\\ \Im\{f(\mathbf{z}_{1}(t))\}\\ \vdots\\ \Im\{f(\mathbf{z}_{M}(t))\}\\ \end{bmatrix}, (7)

where {}\Re\{\cdot\} and {}\Im\{\cdot\} denote the real and imaginary components of {}\{\cdot\} and the subscript denotes the antenna element from which the sample was collected at time tt.

Using a set of training data for which we have both the linear and nonlinear representation of the received signal, 𝐳(t)\mathbf{z}(t) and f(𝐳(t))tf(\mathbf{z}(t))\hskip 2.84526pt\forall\hskip 2.84526ptt, we measure the Euclidean distance between each time sample, i.e., between 𝐳(t)\mathbf{z}(t) and f(𝐳(t))f(\mathbf{z}(t)), and populate a one-hot label vector 𝐲(t)2×1\mathbf{y}(t)\in\mathbb{R}^{2\times 1} for each sample. Specifically, for each time sample, tt, we compute

dt=𝐳(t)f(𝐳(t))2,d_{t}=||\mathbf{z}(t)-f(\mathbf{z}(t))||_{2}, (8)

with the corresponding label determined as

𝐲(t)={[1,0]T,dtdT[0,1]T,dt>dT,\mathbf{y}(t)=\begin{cases}[1,0]^{\text{T}},&d_{t}\leq d_{\text{T}}\\ [0,1]^{\text{T}},&d_{t}>d_{\text{T}}\end{cases}, (9)

where dTd_{\text{T}} is an empirically determined threshold from the training data, and 𝐲(t)\mathbf{y}(t) indicates whether the corresponding time sample is excessively nonlinear.

Next, we train a DNN, ζ:2M×12×1\zeta:\mathbb{R}^{2M\times 1}\rightarrow\mathbb{R}^{2\times 1}, using f(𝐳(t))f(\mathbf{z}(t)) and 𝐲(t)t\mathbf{y}(t)\hskip 2.84526pt\forall\hskip 2.84526ptt to classify the time samples, f(𝐳(t))f(\mathbf{z}(t)), that are highly distorted from their linear counterparts. Note that the input to the DNN is f(𝐳(t))f(\mathbf{z}(t)) since, during test time, we will only receive the nonlinear signal, and thus need to classify its time samples that have contributed to the nonlinearity to the greatest extent. We consider a pp-layered DNN, consisting of AA units each, with a softmax output layer yielding the DNN prediction [𝐲^1(t),𝐲^2(t)][\hat{\mathbf{y}}_{1}(t),\hat{\mathbf{y}}_{2}(t)], where the input is predicted to be nonlinear if 𝐲^1(t)<𝐲^2(t)\hat{\mathbf{y}}_{1}(t)<\hat{\mathbf{y}}_{2}(t). In place of a two-dimensional softmax output, the DNN could output a one-dimensional sigmoidal value. However, we empirically found that the former setup yields a more accurate nonlinear sample classifier.

Using f(𝐳(t))f(\mathbf{z}(t)) and 𝐲(t)t\mathbf{y}(t)\hskip 2.84526pt\forall\hskip 2.84526ptt as training samples, the DNN minimizes the mean squared error loss function. At test time, each column (i.e., time sample) is forward propagated through the DNN, and time samples classified as highly distorted due to the nonlinearity are deleted. The edited matrices of f(𝐙old)f(\mathbf{Z}_{\text{old}}) and f(𝐙new)f(\mathbf{Z}_{\text{new}}), denoted as 𝐙old\mathbf{Z}_{\text{old}}^{\prime} and 𝐙new\mathbf{Z}_{\text{new}}^{\prime}, respectively, correspond to f(𝐙old)f(\mathbf{Z}_{\text{old}}) and f(𝐙new)f(\mathbf{Z}_{\text{new}}) after their time samples that were classified as nonlinear by ζ\zeta were removed. 𝐙old\mathbf{Z}_{\text{old}}^{\prime} and 𝐙new\mathbf{Z}_{\text{new}}^{\prime} are then used to compute the GLRT test statistic in (5).

II-E Complexity Analysis

Here, we analyze the computational complexity of our proposed framework during inference. The GLRT statistic in (5) consists of 22 M×kM\times k matrix Hermitian operations and 22 M×kM\times k matrix multiplications followed by an M×MM\times M matrix inverse and an M×MM\times M matrix multiplication. Thus, the computational complexity of the GLRT is 𝒪(k2+M2)\mathcal{O}(k^{2}+M^{2}). Since the DNN pre-pended to the GLRT in our proposed framework consists of pp dense layers with AA units each, it adds an additional complexity of 𝒪(2(p+1)A3)\mathcal{O}(2(p+1)A^{3}). This yields a total time complexity, for our method, of 𝒪(pA3+k2+M2)\mathcal{O}(pA^{3}+k^{2}+M^{2}). During inference, our method requires roughly 6060 ms to compute the test statistic in (5) for the simulation environment considered in Sec. III. Thus, despite the overhead incurred from the DNN, our method remains computationally feasible for practical use.

III Simulations

III-A Signal Generation

For our empirical evaluation, we generate an interference signal, a1a_{1}, and a SOI, a2a_{2}. Each signal contains a sequence of random bits, is modulated according to the BPSK constellation, and transmitted with a carrier frequency of fc=10f_{c}=10 MHz and a bandwidth of 11 MHz. At the receiver, we use M=4M=4 antenna elements and consider a received signal with an observation window of L=2000L=2000 time samples. In addition, we use k=48k=48 time samples to construct each initial instance of 𝐙old\mathbf{Z}_{\text{old}} and 𝐙new\mathbf{Z}_{\text{new}}. Here, we consider two setups by scaling the AWGN: (a) an interference to noise ratio (INR) of 40 dB and a signal to noise ratio (SNR) of 15 dB (corresponding to perhaps an RF application); and (b) an INR of 57 dB and an SNR of 37 dB (corresponding to perhaps an audio application). The large power difference between the SOI and interference introduces an additional challenge for detection, which is achievable in the linear case but much more challenging in nonlinearity.

In our analysis, we use the hyperbolic tangent function as the induced nonlinearity, i.e., we define f(𝐙)=tanh(𝐙)f(\mathbf{Z})=\text{tanh}(\mathbf{Z}). This function is a representative model for hardware nonlinearities, since it consists of a linear and a nonlinear region, thus reflecting the effect of amplifiers when only certain received samples experience nonlinear saturation [14].

III-B DNN Implementation

The DNN is trained on a generated received signal 𝐙Train\mathbf{Z}_{\text{Train}} and its nonlinear counterpart f(𝐙Train)f(\mathbf{Z}_{\text{Train}}) to identify excessively nonlinear signals based on the method in Sec. II-D. The employed DNN contains p=3p=3 layers with A=10A=10 hyperbolic tangent units each. During training, we use a batch size of 64, a learning rate of 0.0010.001, and 500 epochs with a patience on the training loss of 25 epochs as the stopping criterion.

III-C Detector Performance

Refer to caption
Refer to caption
Figure 2: GLRT performance of a SOI turning on at T0=1000T_{0}=1000 in setups (a) (top) and (b) (bottom). The contrast peak, produced by the SOI turning on at T0T_{0} in the ideal case, is not distinctive in the nonlinear case but reappears after using DNN pre-processing.
Refer to caption
Refer to caption
Figure 3: Detection performance in each setups (a) (top) and (b) (bottom) of the theoretical signal, NL signal, linear and NL signal with our proposed DNN pre-processing, and two baselines. Our methodology significantly improves the detection performance under nonlinearity in comparison to the GLRT operating solely on the NL signal and over the considered baselines.

We evaluate our methodology on f(𝐙Test)f(\mathbf{Z}_{\text{Test}}), and in addition, we generate the linear counterpart, 𝐙Test\mathbf{Z}_{\text{Test}}, to compare the nonlinear and linear detection performance. During implementation, we scale the nonlinearity based on a factor α\alpha as f(α𝐙)/αf(\mathbf{\alpha Z})/\alpha to achieve a desirable nonlinearity in each considered setup. Fig. 2 shows the GLRT test statistic on (i) a theoretical (linear) signal, (ii) a nonlinear signal, and (iii) the same nonlinear signal after DNN pre-processing for setups (a) and (b). The SOI turns on at T0=1000T_{0}=1000. As shown by the metric on the linear signal, a distinctive contrast peak is captured at the time in which the SOI turns on in the received transmission. When the signal undergoes a nonlinearity, and f(𝐙Test)f(\mathbf{Z}_{\text{Test}}) is received without DNN pre-processing, several false alarm peaks are produced at multiple incorrect time samples, and no distinctive contrast peak is captured at T0T_{0} when the SOI turns on. When our proposed methodology is applied to the nonlinear signal, however, the GLRT test statistic produces a distinctive contrast peak correctly at T0=L/2T_{0}=L/2, albeit to somewhat of a lesser degree than in the theoretical case. This behavior stems from the remaining samples in 𝐙old\mathbf{Z}_{\text{old}}^{\prime} and 𝐙new\mathbf{Z}_{\text{new}}^{\prime} still containing nonlinear samples. Using the curves in Fig. 2, we vary the detection threshold and measure the resulting false positive versus true positive detection rates to obtain the performance curves shown next in Fig. 3.

Fig. 3 shows the receiver operating characteristic (ROC) detection performance curves of the GLRT for each setup. In addition, we compare our proposed method to two baselines: (i) end-to-end deep learning (EtE DL) [15], where a deep learning classifier is trained on both linear and nonlinear signals to perform detection directly in place of the GLRT; and (ii) data driven (DD) pre-processing [16], where the received nonlinear signal is first pre-processed in an attempt to recover the linear signal prior to applying the GLRT. In our adoption of [16], we train an autoencoder (AE) to map nonlinear signals (taken as input to the AE) to their linear counterparts (given as output by the AE). The GLRT test statistic in (5) is then applied on the AE output. Finally, for completion, we also employ our proposed DNN pre-processing methodology on a linear signal to show its effect when it is applied to a signal that does not suffer from nonlinear distortions.

From Fig. 3, we see that our method is able to improve the area under the curve (AUC) of the ROC curve in both cases over using the nonlinear signal without DNN pre-processing and the other considered baselines. The lower AUC, in comparison to the theoretical linear case, is attributed to the increased magnitudes of the false positive contrast peaks produced by the GLRT in Fig. 2. In particular, we note that our method significantly improves the true positive detection rate at low false alarm levels, which is often desirable in the context of wireless signal detection. The nearly random performance of the EtE DL classifier is attributed to the large power difference between the interference and SOI, resulting in a lack of salient features for the classifier to perform detection on. Similarly, DD pre-processing faces challenges in the training step since the same nonlinear clipped time sample corresponds to several linear time samples (i.e., a one-to-many mapping), thus resulting in poorly trained AE models. Our proposed method avoids the sensitivity of mapping non-linear time samples to an approximation of their linear counterparts by instead classifying and removing nonlinear samples that degrade detection performance and then applying the GLRT. Finally, we see that applying our proposed framework to a linear signal only slightly degrades performance in comparison to the theoretical linear performance. However, since we can deduce when a received signal has undergone non-linear distortions from hardware constraints, we can employ our proposed framework on non-linear signals while resorting to the classical GLRT on linear observations.

IV Conclusion and Future Work

In this work, we proposed prepending a GLRT detector with a DNN in order to perform reliable detection in the presence of nonlinear distortions. We showed that our method was capable of performing accurate detection despite the nonlinearity, whereas nonlinear signals without any pre-processing were unable to be used for detection using the GLRT. In future work, we anticipate exploring additional nonlinear environments that our framework could be applied to, including the extended target case [17]. In addition, we expect to expand this work to dynamic environments where the direction of the interference or SOI may not be stationary and require the utilization of the GLRT in an FDA-MIMO [18] framework.

References

  • [1] Y. Li, J. Winters, and N. Sollenberger, “MIMO-OFDM for wireless communications: signal detection with enhanced channel estimation,” IEEE Trans. Commun., vol. 50, no. 9, pp. 1471–1477, 2002.
  • [2] A. Salehi, A. Zaimbashi, and M. Valkama, “Kernelized-likelihood ratio tests for binary phase-shift keying signal detection,” IEEE Trans. Cogn. Commun. Netw., vol. 7, no. 2, pp. 541–552, 2021.
  • [3] H. Tang, L. Chai, and X. Wan, “An augmented generalized likelihood ratio test detector for signal detection in clutter and noise,” IEEE Access, vol. 7, pp. 163 478–163 486, 2019.
  • [4] B. Wu, S. Cheng, and H. Wang, “Clipping effects on channel estimation and signal detection in OFDM,” in Proc. of IEEE PIMRC, vol. 1, 2003, pp. 531–534.
  • [5] E. Kelly, “An adaptive detection algorithm,” IEEE Trans. Aerosp. Electron. Syst., vol. AES-22, no. 2, pp. 115–127, 1986.
  • [6] K. Forsythe, “Utilizing waveform features for adaptive beamforming and direction finding with narrowband signals,” Lincoln Laboratory Journal, vol. 10, no. 2, pp. 99–126, 1997.
  • [7] F. Chapeau-Blondeau, “Nonlinear test statistic to improve signal detection in non-gaussian noise,” IEEE Signal Process. Lett., vol. 7, no. 7, pp. 205–207, 2000.
  • [8] D. Rousseau, G. Anand, and F. Chapeau-Blondeau, “Noise-enhanced nonlinear detector to improve signal detection in non-gaussian noise,” Signal Processing, vol. 86, no. 11, pp. 3456–3465, 2006.
  • [9] J. Tellado, L. Hoo, and J. Cioffi, “Maximum-likelihood detection of nonlinearly distorted multicarrier symbols by iterative decoding,” IEEE Trans. Commun., vol. 51, no. 2, pp. 218–228, 2003.
  • [10] H. Ye, G. Y. Li, and B.-H. Juang, “Power of deep learning for channel estimation and signal detection in OFDM systems,” IEEE Wireless Commun. Lett., vol. 7, no. 1, pp. 114–117, 2018.
  • [11] Y. Xie, X. Liu, K. C. Teh, and Y. L. Guan, “Robust deep learning based end-to-end receiver for OFDM system with non-linear distortion,” IEEE Commun. Lett., pp. 1–1, 2021.
  • [12] R. Sahay, C. G. Brinton, and D. J. Love, “A deep ensemble-based wireless receiver architecture for mitigating adversarial attacks in automatic modulation classification,” IEEE Trans. Cogn. Commun. Netw., vol. 8, no. 1, pp. 71–85, 2022.
  • [13] E. J. Kelly and K. Forsythe, “Adaptive detection and parameter estimation for multidimensional signal models,” in Technical Report, MIT Lincoln Laboratory, 1989.
  • [14] S. R. Abdulridha and F. S. Hasan, “Palm clipping and nonlinear companding techniques based papr reduction in ofdm-dcsk system,” Journal of Engineering and Sustainable Development, vol. 25, no. 4, p. 84–94, Feb. 2022.
  • [15] W. Zhang, M. Feng, M. Krunz, and A. Hossein Yazdani Abyaneh, “Signal detection and classification in shared spectrum: A deep learning approach,” in Proc. of IEEE INFOCOM, 2021, pp. 1–10.
  • [16] A. Sayeed, “Data-driven signal detection and classification,” in in Proc. of IEEE ICASSP, vol. 5, 1997, pp. 3697–3700.
  • [17] M. Beard, S. Reuter, K. Granström, B.-T. Vo, B.-N. Vo, and A. Scheel, “Multiple extended target tracking with labeled random finite sets,” IEEE Trans. Signal Process., vol. 64, no. 7, pp. 1638–1653, 2016.
  • [18] L. Lan, A. Marino, A. Aubry, A. De Maio, G. Liao, J. Xu, and Y. Zhang, “Glrt-based adaptive target detection in fda-mimo radar,” IEEE Trans. Aerosp. Electron. Syst., vol. 57, no. 1, pp. 597–613, 2021.