This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Compressed Sensing Radar Detectors
based on Weighted LASSO

Siqi Na1, Yoshiyuki Kabashima2, Takashi Takahashi2, Tianyao Huang1, Yimin Liu1 and Xiqin Wang1 The work of Tianyao Huang was supported by the National Natural Science Foundation of China under Grant 62171259. Email: [email protected] [email protected] [email protected] {huangtianyao, yiminliu, wangxq_ee}@tsinghua.edu.cn
1 Department of Electronic Engineering, Tsinghua University, Beijing, China
2 Institute for Physics of Intelligence, Department of Physics Graduate School of Science, The University of Tokyo, Japan
Abstract

The compressed sensing (CS) model can represent the signal recovery process of a large number of radar systems. The detection problem of such radar systems has been studied in many pieces of literature through the technology of debiased least absolute shrinkage and selection operator (LASSO). While naive LASSO treats all the entries equally, there are many applications in which prior information varies depending on each entry. Weighted LASSO, in which the weights of the regularization terms are tuned depending on the entry-dependent prior, is proven to be more effective with the prior information by many researchers. In the present paper, existing results obtained by methods of statistical mechanics are utilized to derive the debiased weighted LASSO estimator for randomly constructed row-orthogonal measurement matrices. Based on this estimator, we construct a detector, termed the debiased weighted LASSO detector (DWLD), for CS radar systems and prove its advantages. The threshold of this detector can be calculated by false alarm rate, which yields better detection performance than naive weighted LASSO detector (NWLD) under the Neyman-Pearson principle. The improvement of the detection performance brought by tuning weights is demonstrated by numerical experiments. With the same false alarm rate, the detection probability of DWLD is obviously higher than those of NWLD and the debiased (non-weighted) LASSO detector (DLD).

Index Terms:
Compressed sensing, radar detection, LASSO, weighted LASSO, row-orthogonal matrix, replica method, statistical mechanics.

I Introduction

The signal recovery process of radar systems can be represented by a simple linear regression model as follow

𝒚=𝑨𝒙0+𝝃,{\bm{y}}={\bm{A}}{\bm{x}}_{0}+{\bm{\xi}}, (1)

where 𝑨M×N{\bm{A}}\in{\mathbb{C}}^{M\times N} is called measurement matrix, which contains information of the radar transmit waveform. Noise vector 𝝃M{\bm{\xi}}\in{\mathbb{C}}^{M} is usually regarded as additive white Gaussian noise (AWGN) with i.i.d. components ξi𝒞𝒩(0,σ2)\xi_{i}\sim{\cal{CN}}\left({0,\sigma^{2}}\right). Usually, one aims to recover 𝒙0N{\bm{x}}_{0}\in{\mathbb{C}}^{N}, which contains information about the radar observation scene, from the received signal 𝒚M{\bm{y}}\in{\mathbb{C}}^{M}. The increasing development of compressed sensing (CS) technology in recent years has resulted in many CS radar systems, in which signal detection from fewer measurements (M<NM<N) is required. In addition, we find that a number of radar emission waveforms conform to the row-orthogonal assumption, such as partial observation of a pulse Doppler radar system [1], frequency agile radar system [2, 3], sub-Nyquist radar systems [4, 5]. We here defined matrix AA as “row-orthogonal” when it is constructed randomly so as to satisfy AiAjH=δijA_{i\cdot}A_{j\cdot}^{H}=\delta_{ij} 1i,jM1\leq i,j\leq M, where δij=1\delta_{ij}=1 for i=ji=j and 0, otherwise.

Based on results in literature, if 𝒙0{\bm{x}}_{0} is sparse enough, it can be precisely recovered from 𝒚{\bm{y}} in a noise-free scene. That is, the upper bound of kk, which is the number of non-zero entries in 𝒙0{\bm{x}}_{0}, for the perfect recovery is strictly limited by some properties of measurement matrix 𝑨{\bm{A}} such as the compression rate γ=M/N\gamma=M/N. To solve the underdetermined recovery problem (1), CS methods are usually applied, in which the least absolute shrinkage and selection operator (LASSO) [6] is a frequently used technique whose behavior has been studied in much literature [7, 8, 9, 10, 11, 12, 13].

The complex-valued LASSO estimator 𝒙^LASSO\hat{\bm{x}}^{\rm{LASSO}} is given by

𝒙^LASSO=argmin𝒙{12𝒚𝑨𝒙22+λ𝒙1},\hat{\bm{x}}^{\rm{LASSO}}=\mathop{\arg\min}\limits_{{\bm{x}}}\left\{\frac{1}{2}{\left\|{\bm{y}}-{\bm{A}}{\bm{x}}\right\|_{2}^{2}}+\lambda{\left\|{\bm{x}}\right\|_{1}}\right\}, (2)

where

𝒙1=.i=1N(Re(xi))2+(Im(xi))2.{\left\|{\bm{x}}\right\|_{1}}\buildrel\textstyle.\over{=}\sum\limits_{i=1}^{N}{\sqrt{\left({\rm{Re}}\left(x_{i}\right)\right)^{2}+\left({\rm{Im}}\left(x_{i}\right)\right)^{2}}}. (3)

Based on debiased LASSO, which is also known as desparsified LASSO in the context of statistics, given by

𝒙^d=𝒙^LASSO+1Λ𝑨H(𝒚𝑨𝒙^LASSO),{\hat{\bm{x}}}^{\rm{d}}={\hat{\bm{x}}}^{\rm{LASSO}}+\frac{1}{\Lambda}{\bm{A}}^{H}({\bm{y}}-{\bm{A}}{\hat{\bm{x}}}^{\rm{LASSO}}), (4)

where Λ>0\Lambda>0 is the debiased coefficient computed from known variables, there are also some researches [14, 15, 16, 17] on the following detection problems

{0,i:x0,i=0,1,i:x0,i0,\left\{{\begin{array}[]{*{20}{l}}{{{\cal H}_{0,i}}:{x_{0,i}}=0},\\ {{{\cal H}_{1,i}}:{x_{0,i}}\neq 0},\end{array}}\right. (5)

for i=1,2,,Ni=1,2,\ldots,N. As mentioned above, the indices of the non-zero elements in 𝒙0{\bm{x}}_{0} indicate information about the target, and determining whether each element in 𝒙0{\bm{x}}_{0} is non-zero or not can inform us about the existence and location of the targets. Consequently, dealing with problems (5) means to detect targets element-wisely, which is very critical for modern radar systems.

There are also some applications for radar systems that we may acquire some prior information about 𝒙0{\bm{x}}_{0}, which refers to the support set or the probability that each element in 𝒙0{\bm{x}}_{0} is non-zero. For instance, the multi-frame observations with tracking [18] brings us the prediction of the target’s position in the next frame. A weighted version of LASSO, weighted LASSO (or weighted 1\ell_{1} minimization for noise-free scenario), has been studied to exploit prior information in the process of signal recovery [19, 20, 21, 22], which is given by

𝒙^WL=argmin𝒙{12𝒚𝑨𝒙22+i=1Nλixi},\hat{\bm{x}}^{\rm{WL}}=\mathop{\arg\min}\limits_{{\bm{x}}}\left\{\frac{1}{2}{\left\|{\bm{y}}-{\bm{A}}{\bm{x}}\right\|_{2}^{2}}+\sum\limits_{i=1}^{N}\lambda_{i}\|x_{i}\|\right\}, (6)

where 𝝀=.[λ1,λ2,,λN]T{\bm{\lambda}}\buildrel\textstyle.\over{=}[\lambda_{1},\lambda_{2},\ldots,\lambda_{N}]^{T} is the vector of weights. Researches mentioned above are mainly based on restricted isometry property (RIP) of measurement matrix 𝑨\bm{A} and provide results on the upper bound of 𝒙^WL𝒙02\|\hat{\bm{x}}^{\rm{WL}}-{\bm{x}}_{0}\|_{2}. Reference [23] shows the advantages of proper weighting according to prior information for noise-free cases.

As for the debiased strategy, we are interested in the debiased weighted LASSO estimator, given by

𝒙^d=𝒙^WL+1Λ𝑨H(𝒚𝑨𝒙^WL).{\hat{\bm{x}}}^{\rm{d}}={\hat{\bm{x}}}^{\rm{WL}}+\frac{1}{\Lambda}{\bm{A}}^{H}({\bm{y}}-{\bm{A}}{\hat{\bm{x}}}^{\rm{WL}}). (7)

With results in [24] and some variable substitution, one can obtain a debiased weighted LASSO estimator under random Gaussian design, that is, the elements of 𝑨\bm{A} are i.i.d. drawn from a Gaussian distribution. Our goal is to derive a debiased weighted LASSO estimator, particularly for complex-valued row-orthogonal measurement matrix, and construct its related detector, which we term the debiased weighted LASSO detector (DWLD).

In this paper, we provide the framework of DWLD and analyze its detection performance by comparing DWLD with the naive weighted LASSO detector (NWLD), which is defined as a set of detectors rejecting 0,i{\cal H}_{0,i} if |x^iWL||{\hat{x}}^{\rm{WL}}_{i}| is greater than a given threshold. When the measurement matrix is complex-valued row-orthogonal, the concrete implementation of DWLD is presented employing results obtained by methods of statistical mechanics [16, 17]. Merits of DWLD are twofold: the threshold of DWLD can be analytically calculated by the probability of false alarm, which enables DWLD to achieve a better detection performance than NWLD under the Neyman-Pearson principle. We propose a method to optimize the setting of weights 𝝀{\bm{\lambda}} with the goal of improving the detection performance. We also demonstrate that proper weighting can improve detection performance than the debiased LASSO detector (DLD) [17] through numerical experiments.

The present paper is organized as follows. In Section II, we implement DWLD for compressed sensing radar system with prior information providing theoretical consideration on its detection performance. The approach to optimize the weights, in order to improve detection performance, is also proposed. Section III provides some numerical results of the detection performance of DWLD. We conclude the paper in Section IV.

Throughout the paper, we use aa, 𝒂\bm{a} and 𝑨\bm{A} as a number, a vector and a matrix, respectively. For a set SS, #S\#S denotes its cardinality. Function δ()\delta(\cdot) is the Dirac’s delta function and Θ(x)\Theta(x) is Heaviside’s step function. The operators ()T(\cdot)^{T}, ()(\cdot)^{*} and ()H(\cdot)^{H} represent the transpose, conjugate and conjugate transpose of a component, respectively. The indicator function is of a subset AA of a set XX is presented by 𝟏A(x){\bm{1}}_{A}(x).

II Design and analysis of debiased weighted LASSO detector

In this section, we first provide some metrics with regard to detection performance in Section II-A. Then the framework of DWLD together with NWLD is introduced in Section II-B. We also give a concrete implementation of DWLD under row-orthogonal design, including the procedure for calculating thresholds based on false alarm rates. The advantages of DWLD are elaborated mainly by comparing with NWLD in Section II-C. At last, the method to optimize the weights 𝝀\bm{\lambda} is proposed in Section II-D.

II-A Preliminaries

For the NN detection problems described in (5), we here define the following performance metrics. Suppose we perform the tests for TT trials. Denote by

φi(t)={1,if detector reject 0,i in t-th trial;0,otherwise,{\varphi_{i}^{(t)}}=\left\{{\begin{array}[]{*{20}{c}}{1,}&{{\text{if detector reject }}{{\cal{H}}_{0,i}}{\text{ in $t$-th trial}};}\\ {0,}&{{\text{otherwise,}}}\end{array}}\right. (8)

for i=1,,Ni=1,\ldots,N and t=1,,Tt=1,\ldots,T. Define the probability of false alarm Pfa,iP_{fa,i} for the ii-th problem or entry of 𝒙0{\bm{x}}_{0} as

Pfa,i=limTt=1Tφi(t)𝟏Stc(i)t=1T𝟏Stc(i),P_{fa,i}=\mathop{\lim}\limits_{T\to\infty}\frac{\sum\limits_{t=1}^{T}{\varphi_{i}^{(t)}\cdot{\bm{1}}_{S_{t}^{c}}(i)}}{\sum\limits_{t=1}^{T}{{\bm{1}}_{S_{t}^{c}}(i)}}, (9)

and the probability of detection Pd,iP_{d,i} as

Pd,i=limTt=1Tφi(t)𝟏St(i)t=1T𝟏St(i),P_{d,i}=\mathop{\lim}\limits_{T\to\infty}\frac{\sum\limits_{t=1}^{T}{\varphi_{i}^{(t)}\cdot{\bm{1}}_{S_{t}}(i)}}{\sum\limits_{t=1}^{T}{{\bm{1}}_{S_{t}}(i)}}, (10)

where St={i:x0,i(t)0}S_{t}=\left\{i:x_{0,i}^{(t)}\neq 0\right\} is the support set of 𝒙0(t){\bm{x}}_{0}^{(t)} with Stc={1,,N}\StS_{t}^{c}=\{1,\ldots,N\}\backslash S_{t}.

The prior information 𝒑=[p1,p2,,pN]T{\bm{p}}=[p_{1},p_{2},\ldots,p_{N}]^{T} of 𝒙0{\bm{x}}_{0} is defined as the probability that each entry of 𝒙0{\bm{x}}_{0} is non-zero, such that

P(𝒙0)=i=1N[(1pi)δ(x0,i)+piQi(x0,i)],P({\bm{x}}_{0})=\prod\limits_{i=1}^{N}\left[(1-p_{i})\delta(x_{0,i})+p_{i}Q_{i}(x_{0,i})\right], (11)

where Qi(x0,i)Q_{i}(x_{0,i}) is the distribution of non-zero entry of 𝒙0{\bm{x}}_{0}.

Since NN detection problems are processed simultaneously, we define the total false alarm rate P¯fa\bar{P}_{fa} as follow,

P¯fa=limTt=1Ti=1Nφi(t)𝟏Stc(i)t=1Ti=1N𝟏Stc(i),{\bar{P}_{fa}}=\mathop{\lim}\limits_{T\to\infty}\frac{\sum\limits_{t=1}^{T}\sum\limits_{i=1}^{N}{\varphi_{i}^{(t)}\cdot{\bm{1}}_{S_{t}^{c}}(i)}}{\sum\limits_{t=1}^{T}\sum\limits_{i=1}^{N}{{\bm{1}}_{S_{t}^{c}}(i)}}, (12)

and similarly the total detection rate P¯d\bar{P}_{d}

P¯d=limTt=1Ti=1Nφi(t)𝟏St(i)t=1Ti=1N𝟏St(i).{\bar{P}_{d}}=\mathop{\lim}\limits_{T\to\infty}\frac{\sum\limits_{t=1}^{T}\sum\limits_{i=1}^{N}{\varphi_{i}^{(t)}\cdot{\bm{1}}_{S_{t}}(i)}}{\sum\limits_{t=1}^{T}\sum\limits_{i=1}^{N}{{\bm{1}}_{S_{t}}(i)}}. (13)

With the definition of prior information 𝒑\bm{p} above, one can say that

pi=limT1Tt=1T𝟏St(i).p_{i}=\mathop{\lim}\limits_{T\to\infty}\frac{1}{T}\sum\limits_{t=1}^{T}{{\bm{1}}_{S_{t}}(i)}. (14)

Therefore,

P¯fa=i=1N(1pi)Pfa,i/i=1N(1pi),\displaystyle{\bar{P}_{fa}}=\sum\limits_{i=1}^{N}(1-p_{i})P_{fa,i}/\sum\limits_{i=1}^{N}(1-p_{i}), (15)
P¯d=i=1NpiPd,i/i=1Npi.\displaystyle{\bar{P}_{d}}=\sum\limits_{i=1}^{N}p_{i}P_{d,i}/\sum\limits_{i=1}^{N}p_{i}. (16)

Note that if we take the limit NN\to\infty, these definition will be the same as that in [17], with (12) and (13) considering prior information.

II-B Design of the debiased weighted LASSO detector

Refer to caption
Figure 1: Frameworks of two detectors for detection problem (5): (a) naive weighted LASSO detector (NWLD); (b) debiased weighted LASSO detector (DWLD).

Similarly to [17], we demonstrate and compare DWLD with NWLD. Their frameworks are shown in Fig. 1 (a) and (b). The main difference between the two detectors is in the test statistics and thresholds. The latter performs a “debiased” procedure, which is a linear transform of the weighted LASSO estimator, and applies calculable thresholds with given false alarm rate for each entry. The details of the superiority of DWLD and their proofs will be presented in Sec. II-C.

Due to the random nature of the row-orthogonal matrix 𝑨{\bm{A}}, 𝒙^d{\hat{\bm{x}}}^{\rm{d}} is expected to follow Gaussian distribution with mean 𝒙0{\bm{x}}_{0}. Actually, this is the case as long as 𝑨{\bm{A}} is “right rotation invariant”, which means that the right singular basis of 𝑨{\bm{A}} is regarded as a random sample from the Haar measure of N×NN\times N orthogonal (or unitary) matrices. Further, in such cases, methods of statistical mechanics offer formulas to appropriately construct 𝒙^d{\hat{\bm{x}}}^{\rm{d}} from the (weighted) LASSO estimator and to analytically evaluate the variance of 𝒙^d{\hat{\bm{x}}}^{\rm{d}} depending on the asymptotic eigenvalue distribution ρ𝑱(s)\rho_{\bm{J}}(s) of 𝑱=𝑨T𝑨{\bm{J}}={\bm{A}}^{T}{\bm{A}} (or 𝑱=𝑨H𝑨{\bm{J}}={\bm{A}}^{H}{\bm{A}}) [16, 17]. Utilizing the formulas, we here propose DWLD for complex-valued row-orthogonal observation matrix, whose procedure is shown in Algorithm 1. Due to space limitation, we leave its derivation to [17].

Algorithm 1 DWLD under complex row-orthogonal design
1:𝒚\bm{y}, 𝑨\bm{A}, weights 𝝀{\bm{\lambda}}, probability of false alarm {Pfa,i}i=1N\{P_{fa,i}\}_{i=1}^{N}, variance of noise σ2\sigma^{2}
2:debiased weighted LASSO estimator 𝒙^d{\hat{\bm{x}}}^{\rm{d}}, threshold {κd,i}i=1N\{\kappa_{d,i}\}_{i=1}^{N}
3:Let
𝒙^WL=argmin𝒙{12𝒚𝑨𝒙22+i=1Nλixi}.\hat{\bm{x}}^{\rm{WL}}=\mathop{\arg\min}\limits_{{\bm{x}}}\left\{\frac{1}{2}{\left\|{\bm{y}}-{\bm{A}}{\bm{x}}\right\|_{2}^{2}}+\sum\limits_{i=1}^{N}\lambda_{i}\|x_{i}\|\right\}.
4:The debiased weighted LASSO estimator is obtained from
𝒙^d=𝒙^WL+1ΛCRO𝑨H(𝒚𝑨𝒙^WL),{\hat{\bm{x}}}^{\rm{d}}={\hat{\bm{x}}}^{\rm{WL}}+\frac{1}{\Lambda_{\rm{CRO}}}{\bm{A}}^{H}({\bm{y}}-{\bm{A}}{\hat{\bm{x}}}^{\rm{WL}}), (17)
with ΛCRO\Lambda_{\rm{CRO}} and ρCA\rho_{\rm{CA}} are solved from the following fixed point equation
ΛCRO\displaystyle\Lambda_{\rm{CRO}} =\displaystyle= γρCA1ρCA,\displaystyle\frac{\gamma-\rho_{\rm{CA}}}{1-\rho_{\rm{CA}}},
ρCA\displaystyle\rho_{\rm{CA}} =\displaystyle= 12Ni=1N[(2λiΛCRO|x^iWL|+λi)\displaystyle\frac{1}{2N}\sum\limits_{i=1}^{N}\left[\left(2-\frac{\lambda_{i}}{\Lambda_{\rm{CRO}}\left|{\hat{x}_{i}^{\rm{WL}}}\right|+\lambda_{i}}\right)\right. (18)
Θ(|x^iWL|)].\displaystyle\left.\cdot\Theta\left(\left|{\hat{x}_{i}^{\rm{WL}}}\right|\right)\right].
5:The threshold {κd,i}i=1N\{\kappa_{d,i}\}_{i=1}^{N} is given by
κd,i=σw2lnPfa,i,\kappa_{d,i}=-\sigma_{w}^{2}\ln{P_{fa,i}},
where
σw2=γ(1γ)(γρCA)2RSS¯+σ2,\displaystyle\sigma_{w}^{2}=\frac{\gamma(1-\gamma)}{\left(\gamma-\rho_{\rm{CA}}\right)^{2}}\overline{\rm{RSS}}+\sigma^{2}, (19)
RSS¯=1M𝒚𝑨𝒙^WL22.\displaystyle\overline{\rm{RSS}}=\frac{1}{M}\left\|\bm{y}-\bm{A}{\hat{\bm{x}}}^{\rm{WL}}\right\|_{2}^{2}. (20)

One can deduce results similar to (17) for random Gaussian measurement matrices from some conclusions in [24]. More specifically, with Theorem 10 in [24], the debiased coefficient for random Gaussian design is ΛG=γρa\Lambda_{G}=\gamma-\rho_{a}, where ρa=#{i|x^iWL0}/N\rho_{a}=\#\{i|{\hat{x}_{i}^{\rm{WL}}}\neq 0\}/N denotes the active component density for real-valued situation. We emphasize that the methodologies in [16, 17] not only reproduce the same result for the Gaussian matrices but also can offer the debiased coefficients for general right rotation invariant matrices.

II-C Analysis of detection performance

In general, the advantage of DWLD is that its threshold κd,i\kappa_{d,i} can be calculated analytically from the false alarm rate Pfa,iP_{fa,i}. On the other hand, due to the fact that there is no analytic solution for weighted LASSO, one cannot establish the relationship between the threshold κWL,i\kappa_{WL,i} of NWLD and the false alarm rate Pfa,iP_{fa,i}. We will next introduce the two benefits brought by DWLD in Section II-C1 and II-C2, respectively.

II-C1 Analytical expression of false alarm probability

Controlling the probability of false alarm is very important for radar applications. Under the Neyman-Pearson criterion, the performance of the detectors can be compared only if the upper bound of the false alarm rate is controlled. Therefore, we state that the most significant strength of DWLD is that the analytical relationship between the threshold κd,i\kappa_{d,i} and the false alarm rate Pfa,iP_{fa,i} is available. As shown in [24], for the debiased estimator (7), the empirical distribution of the difference

𝒘=.𝒙^d𝒙0,{\bm{w}}\buildrel\textstyle.\over{=}{\hat{\bm{x}}}^{\rm{d}}-{\bm{x}}_{0}, (21)

converges weakly to a zero-mean Gaussian distribution as NN\to\infty for a specific Λ\Lambda. Denote the sample variance of 𝒘{\bm{w}} by σw2\sigma_{w}^{2}, one can get the following analytical relationship between the probability of false alarm Pfa,iP_{fa,i} and the threshold κd,i\kappa_{d,i} of the detector.

Theorem II.1.

When NN\to\infty, the false alarm probability of ii-th entry Pfa,iP_{fa,i} of DWLD satisfies:

κd,i=σw2lnPfa,i,\kappa_{d,i}=-\sigma_{w}^{2}\ln P_{fa,i}, (22)

where the test is

φi={1,|x^id|2>κd,i;0,otherwise.{\varphi_{i}}=\left\{{\begin{array}[]{*{20}{l}}{1,}&{\left|\hat{x}^{\rm{d}}_{i}\right|^{2}>\kappa_{d,i};}\\ {0,}&{{\text{otherwise.}}}\end{array}}\right. (23)

The readers are referred to [17] for the proof. Due to the fact that almost all of the solutions obtained by CS methods does not have a closed form, neither the distribution of the solutions, it is hard to obtain such conclusion for conventional CS detectors.

II-C2 Better detection performance

Based on the definition and subdifferential properties of a convex function, one can draw the following conclusion.

Theorem II.2.

For the same false alarm probability Pfa,iP_{fa,i}, the detection probability of DWLD Pd,i(1)P_{d,i}^{(1)} is not less than that of NWLD Pd,i(2)P_{d,i}^{(2)}.

The proof of this conclusion is similar as that of Theorem 2.7 in [17], which we refer the readers to. Theorem II.2 guarantees that DWLD has a better detection performance comparing with weighted LASSO detector.

II-D Optimization of weights

One can easily conclude from the previous analysis that the smaller σw2\sigma_{w}^{2} is, the better the detection performance of DWLD can achieve. Given other parameters (e.g. distribution and size of measurement matrix 𝑨{\bm{A}}, variance of the noise σ2\sigma^{2}, distribution of 𝒙0{\bm{x}}_{0}, etc.), σw2\sigma_{w}^{2} can be expressed as a function of weights 𝝀\bm{\lambda} and prior information 𝒑\bm{p}, given by

σw2=f1(𝝀,𝒑).\sigma_{w}^{2}=f_{1}\left({\bm{\lambda}},{\bm{p}}\right). (24)

However, f1f_{1} is unavailable because we do not know 𝒙0{\bm{x}}_{0}. Instead, we can obtain its estimation σ^w2\hat{\sigma}_{w}^{2} by the method developed in [16] and [17], which can be represented by

𝔼𝑨,𝝃,𝒙0(σ^w2)=f2(𝝀,𝒑),{\mathbb{E}}_{{\bm{A}},{\bm{\xi}},{\bm{x}}_{0}}\left(\hat{\sigma}_{w}^{2}\right)=f_{2}\left({\bm{\lambda}},{\bm{p}}\right), (25)

where the specific implementation of f2f_{2} can be obtained from Algorithm 1. Therefore, we can find the optimal weights 𝝀\bm{\lambda} for the prior information 𝒑{\bm{p}} by optimizing f2f_{2}, given by

𝝀=argmin𝝀{f2(𝝀,𝒑)}=.f3(𝒑).{\bm{\lambda}}^{*}=\mathop{\arg\min}\limits_{{\bm{\lambda}}}\left\{f_{2}\left({\bm{\lambda}},{\bm{p}}\right)\right\}\buildrel\textstyle.\over{=}f_{3}({\bm{p}}). (26)

Unfortunately, f2f_{2} dose not have an analytic expression. Employing heuristic algorithms to solve an NN-dimensional parameter optimization problem is not very practical. Therefore, we reduced the degrees of freedom by parameterizing the weight vector 𝝀\bm{\lambda} with a few hyper-parameters, such as a linear model λiλ0αpi\lambda^{*}_{i}\simeq\lambda_{0}-\alpha p_{i}, or an exponential model λiλ0/piα\lambda^{*}_{i}\simeq\lambda_{0}/p_{i}^{\alpha}, and obtained sub-optimal results. Consequently, considering the linear model for example, the optimization problem in (26) becomes

{λ0,α}=argmin{a,b}{f2({abpi}i=1N,𝒑)}.\{\lambda_{0},\alpha\}=\mathop{\arg\min}\limits_{\{a,b\}}\left\{f_{2}\left(\{a-bp_{i}\}_{i=1}^{N},{\bm{p}}\right)\right\}. (27)

The simulation results in Section III will verify the effectiveness of proposed weights optimization approach.

III Numerical Experiments

We will demonstrate the detection performance of DWLD through numerical simulation, mainly under row-orthogonal design. We first verify the ability of DWLD to determine the threshold based on the false alarm rate in Sec. III-B, in which we compare DWLD and NWLD. Then, in Sec. III-C, the comparison between DWLD with DLD under complex row-orthogonal design, which is termed CROD in [17], illustrates that proper setting of weights 𝝀\bm{\lambda} can lead to improved detection performance.

III-A Settings

In all the numerical experiments, we artificially generate the original signal 𝒙0{\bm{x}}_{0}, observation matrix 𝑨{\bm{A}} and the noise 𝝃{\bm{\xi}}. The original signal 𝒙0{\bm{x}}_{0} is generated from the Bernoulli-Gaussian distribution and the entries are independent with prior information 𝒑{\bm{p}}, thus

P(𝒙0)=i=1N[(1pi)δ(x0,i)+piπσx2e|x0,i|2/σx2].P({\bm{x}}_{0})=\prod\limits_{i=1}^{N}\left[(1-p_{i})\delta(x_{0,i})+\frac{p_{i}}{\pi\sigma_{x}^{2}}{\rm{e}}^{-|x_{0,i}|^{2}/\sigma_{x}^{2}}\right]. (28)

The measurement matrix 𝑨{\bm{A}} is decided by the radar system. The entries of the noise 𝝃{\bm{\xi}} are i.i.d. complex Gaussian variables: ξi𝒞𝒩(0,σ2)\xi_{i}\sim{\cal{CN}}(0,\sigma^{2}). We adopt the matched filtering (MF) definition of the signal-to-noise ratio (SNR) [5], such that

SNR=γσx2σ2.{\rm{SNR}}=\frac{\gamma\sigma_{x}^{2}}{\sigma^{2}}. (29)

III-B Comparison of DWLD and NWLD

Refer to caption
Figure 2: Total false alarm rate of DWLD and NWLD vary with SNR for the first prior information setting (30). Here γ=0.5\gamma=0.5.
Refer to caption
Figure 3: Total false alarm rate of DWLD and NWLD vary with SNR for the second prior information setting (31). Here γ=0.5\gamma=0.5.

We apply DWLD and NWLD for solving the sub-Nyquist radar detection problem, in which the observation matrix 𝑨\bm{A} is partial Fourier, and examine their false alarm rate. In the following experiments, we set the length of 𝒙0{\bm{x}}_{0} to N=512N=512 and the variance of noise σ2\sigma^{2} to 0.010.01. The probability of false alarm is set to 0.010.01. All the results are obtained by 10410^{4} Monte-Carol trials.

The result of the first experiment is shown in Fig. 2, where the prior information 𝒑\bm{p} is set as

pi={0.01,1i34N,0.8,34N<iN.{p_{i}}=\left\{{\begin{array}[]{*{20}{c}}{0.01,}&{1\leq i\leq\frac{3}{4}N,}\\ {0.8,}&{\frac{3}{4}N<i\leq N.}\end{array}}\right. (30)

The weight vectors 𝝀1{\bm{\lambda}}_{1} and 𝝀2{\bm{\lambda}}_{2} of the two detectors are set to λ1,i=0.10.1pi\lambda_{1,i}=0.1-0.1p_{i} and λ2,i=0.150.15pi\lambda_{2,i}=0.15-0.15p_{i}, respectively. The threshold of NWLD is set to 0, which corresponds to the naive decision based on the weighted LASSO estimator. From the result we can see that DWLD has a good ability to control the false alarm rate when different parameters vary while NWLD dose not.

The second one sets the prior information 𝒑\bm{p} as

pi={0.05,1i34N,0.6,34N<iN.{p_{i}}=\left\{{\begin{array}[]{*{20}{c}}{0.05,}&{1\leq i\leq\frac{3}{4}N,}\\ {0.6,}&{\frac{3}{4}N<i\leq N.}\end{array}}\right. (31)

The setting of weights and thresholds are the same as the former one. The result is shown in Fig. 3, in which DWLD also controls the false alarm rate completely.

III-C Detection performance of DWLD

Refer to caption
Figure 4: (a) Total false alarm rate and (b) total detection rate of different detectors vary with SNR for the first prior information setting (30). Here γ=0.5\gamma=0.5.
Refer to caption
Figure 5: (a) Total false alarm rate and (b) total detection rate of different detectors vary with SNR for the second prior information setting (31). Here γ=0.5\gamma=0.5.

We then compare the detection performance of DWLD and DLD, and the parameter settings of 𝑨\bm{A}, 𝒙0{\bm{x}}_{0} and 𝝃{\bm{\xi}} are the same as Sec. III-B. The result of the first experiment is shown in Fig. 4, where the prior information 𝒑\bm{p} is setting as (30). We compare the detection performance of DWLD with DLD for two regularization parameter settings. The regularization parameters of DLD λL\lambda_{L} are set to be 0.10.1 and 0.20.2. Meanwhile, the weight vectors 𝝀\bm{\lambda} were optimized to λi=0.10000.1002pi\lambda_{i}=0.1000-0.1002p_{i} for the linear model and λi=0.0212/pi0.3470\lambda_{i}=0.0212/p_{i}^{0.3470} for the exponential model for 𝒑\bm{p} of (30). We demonstrate how the detection performance of these detectors varies with SNR. The results suggest that both detectors can maintain the false alarm rate with different SNR, and the total detection rate of DWLD is higher than DLD.

The second one sets the prior information 𝒑\bm{p} as (31). The regularization parameters of DLD λL\lambda_{L} are set to be 0.10.1 and 0.20.2, and the optimal weight vectors 𝝀\bm{\lambda} were λi=0.09630.1038pi\lambda_{i}=0.0963-0.1038p_{i} for the linear model and λi=0.0281/pi0.3594\lambda_{i}=0.0281/p_{i}^{0.3594} for the exponential model. The result is shown in Fig. 5, which indicates that DWLD also presents better detection performance than DLD.

IV Conclusion

The present paper introduces the design of the detector based on debiased weighted LASSO estimator for CS radar systems. The detection performance is analyzed theoretically and proved to be better than the naive one. Aiming at improving the detection performance, we propose an approach to optimize the weights 𝝀\bm{\lambda}. By comparing DWLD with DLD, numerical results show that proper design of weights improves the detection performance.

References

  • [1] L. Xiao, Y. Liu, T. Huang, X. Liu, and X. Wang, “Distributed target detection with partial observation,” IEEE Transactions on Signal Processing, vol. 66, no. 6, pp. 1551–1565, 2018.
  • [2] T. Huang, Y. Liu, X. Xu, Y. C. Eldar, and X. Wang, “Analysis of frequency agile radar via compressed sensing,” IEEE Transactions on Signal Processing, vol. 66, no. 23, pp. 6228–6240, 2018.
  • [3] L. Wang, T. Huang, and Y. Liu, “Randomized stepped frequency radars exploiting block sparsity of extended targets: A theoretical analysis,” IEEE Transactions on Signal Processing, vol. 69, pp. 1378–1393, 2021.
  • [4] D. Cohen and Y. C. Eldar, “Sub-nyquist radar systems: Temporal, spectral, and spatial compression,” IEEE Signal Processing Magazine, vol. 35, no. 6, pp. 35–58, 2018.
  • [5] S. Na, K. V. Mishra, Y. Liu, Y. C. Eldar, and X. Wang, “Tendsur: Tensor-based 4d sub-nyquist radar,” IEEE Signal Processing Letters, vol. 26, no. 2, pp. 237–241, 2018.
  • [6] R. Tibshirani, “Regression shrinkage and selection via the lasso,” Journal of the Royal Statistical Society: Series B (Methodological), vol. 58, no. 1, pp. 267–288, 1996.
  • [7] E. Greenshtein and Y. Ritov, “Persistence in high-dimensional linear predictor selection and the virtue of overparametrization,” Bernoulli, vol. 10, no. 6, pp. 971–988, 2004.
  • [8] P. J. Bickel, Y. Ritov, and A. B. Tsybakov, “Simultaneous analysis of lasso and dantzig selector,” The Annals of statistics, vol. 37, no. 4, pp. 1705–1732, 2009.
  • [9] E. Candes and T. Tao, “The dantzig selector: Statistical estimation when p is much larger than n,” The annals of Statistics, vol. 35, no. 6, pp. 2313–2351, 2007.
  • [10] G. Raskutti, M. J. Wainwright, and B. Yu, “Minimax rates of estimation for high-dimensional linear regression over q\ell_{q}-balls,” IEEE transactions on information theory, vol. 57, no. 10, pp. 6976–6994, 2011.
  • [11] N. Meinshausen and P. Bühlmann, “High-dimensional graphs and variable selection with the lasso,” The annals of statistics, vol. 34, no. 3, pp. 1436–1462, 2006.
  • [12] P. Zhao and B. Yu, “On model selection consistency of lasso,” The Journal of Machine Learning Research, vol. 7, pp. 2541–2563, 2006.
  • [13] M. J. Wainwright, “Sharp thresholds for high-dimensional and noisy sparsity recovery using 1\ell_{1}-constrained quadratic programming (lasso),” IEEE transactions on information theory, vol. 55, no. 5, pp. 2183–2202, 2009.
  • [14] A. Javanmard and A. Montanari, “Hypothesis testing in high-dimensional regression under the gaussian random design model: Asymptotic theory,” IEEE Transactions on Information Theory, vol. 60, no. 10, pp. 6522–6554, 2014.
  • [15] L. Anitori, A. Maleki, M. Otten, R. G. Baraniuk, and P. Hoogeboom, “Design and analysis of compressed sensing radar detectors,” IEEE Transactions on Signal Processing, vol. 61, no. 4, pp. 813–827, 2012.
  • [16] T. Takahashi and Y. Kabashima, “A statistical mechanics approach to de-biasing and uncertainty estimation in lasso for random measurements,” Journal of Statistical Mechanics: Theory and Experiment, vol. 2018, no. 7, p. 073405, 2018.
  • [17] S. Na, T. Huang, Y. Liu, T. Takahashi, Y. Kabashima, and X. Wang, “Compressed sensing radar detectors under the row-orthogonal design model: a statistical mechanics perspective,” arXiv preprint arXiv:2209.15273, 2022.
  • [18] S. Na, T. Huang, Y. Liu, and X. Wang, “Track-before-detect for sub-nyquist radar,” in ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).   IEEE, 2020, pp. 6029–6033.
  • [19] N. Vaswani and W. Lu, “Modified-cs: Modifying compressive sensing for problems with partially known support,” IEEE Transactions on Signal Processing, vol. 58, no. 9, pp. 4595–4607, 2010.
  • [20] D. Liang and L. Ying, “Compressed-sensing dynamic mr imaging with partially known support,” in 2010 Annual International Conference of the IEEE Engineering in Medicine and Biology.   IEEE, 2010, pp. 2829–2832.
  • [21] D. Needell, R. Saab, and T. Woolf, “Weighted-minimization for sparse recovery under arbitrary prior information,” Information and Inference: A Journal of the IMA, vol. 6, no. 3, pp. 284–309, 2017.
  • [22] L. Lian, A. Liu, and V. K. Lau, “Weighted lasso for sparse recovery with statistical prior support information,” IEEE Transactions on Signal Processing, vol. 66, no. 6, pp. 1607–1618, 2018.
  • [23] T. Tanaka and J. Raymond, “Optimal incorporation of sparsity information by weighted 1\ell_{1} optimization,” in 2010 IEEE International Symposium on Information Theory.   IEEE, 2010, pp. 1598–1602.
  • [24] M. Celentano, A. Montanari, and Y. Wei, “The lasso with general gaussian designs with applications to hypothesis testing,” arXiv preprint arXiv:2007.13716, 2020.