This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Robust lEarned Shrinkage-Thresholding (REST): Robust unrolling for sparse recovery

Wei Pu, Chao Zhou, Yonina C. Eldar and Miguel R.D. Rodrigues W. Pu, C. Zhou and M. Rodrigues are with the Department of Electronic and Electrical Engineering, University College London, UK. Y. C. Eldar is with the Weizmann Institute of Science, Rehovot, Israel. This work was supported both by The Alan Turing Institute and Weizmann – UK Making Connections Programme (Ref. 129589)
Abstract

In this paper, we consider deep neural networks for solving inverse problems that are robust to forward model mis-specifications. Specifically, we treat sensing problems with model mismatch where one wishes to recover a sparse high-dimensional vector from low-dimensional observations subject to uncertainty in the measurement operator. We then design a new robust deep neural network architecture by applying algorithm unfolding techniques to a robust version of the underlying recovery problem. Our proposed network - named Robust lEarned Shrinkage-Thresholding (REST) - exhibits an additional normalization processing compared to Learned Iterative Shrinkage-Thresholding Algorithm (LISTA), leading to reliable recovery of the signal under sample-wise varying model mismatch. The proposed REST network is shown to outperform state-of-the-art model-based and data-driven algorithms in both compressive sensing and radar imaging problems wherein model mismatch is taken into consideration.

Index Terms:
Inverse Problems, Compressive Sensing Problems, Model Mismatch, Robustness, Deep Learning

I Introduction

Inverse problems, involving the recovery of a high-dimensional vector of interest from a low-dimensional observation vector, arise in various signal and image processing applications such as compressive sensing[1, 2, 9], synthetic aperture radar (SAR)[4, 5], medical imaging[6, 7, 8], and many more. Three major classes of approaches have been developed over the years to solve such inverse problems. Classical model-based techniques leverage knowledge of the underlying linear model to solve inverse problems via the formulation of optimization problems that include two terms in the objective: (1) a data fidelity term and (2) a data regularization term. The fidelity term encourages the solution to be consistent with the observations whereas the regularization encourages solutions that conform to a certain postulated data prior. For example, variational methods use a regularizer that promotes smoothness of the solutions [10, 11] whereas sparsity-driven methods use regularizers that promote sparse solutions in some transform domain[9, 12].

More recent data-driven techniques “solve” an inverse problem by using powerful neural networks that learn how to map the model output to the model input based on a number of input-output examples[13, 14, 15, 16]. These approaches offer various advantages over model based ones including higher performance and lower complexity. 111Data-driven approaches shift the computational burden from the testing phase to the training phase. A neural network – once optimized – can deliver an estimate of the target quantity more rapidly than a classical optimization based method. However, state-of-the-art data-driven approaches such as deep learning also exhibit various disadvantages: (i) they are often based on heuristic network designs requiring hand-crafted trial-and-error (ii) they typically lack interpretability, including physical meaning; and importantly (iii) they require rich datasets to learn the inverse operation mapping observations to quantities of interest that are not always available.

Finally, in view of the fact that the underlying inverse problem model is often (approximately) known in various settings, there has been increased interest in model-aware data-driven approaches to inverse problems. A popular model-aware data-driven approach relies on algorithm unfolding or unrolling[17, 18, 19, 20, 21]. It involves (i) the formulation of an optimization problem allowing to recover the quantities of interest from the observed quantities; (ii) the identification of an iterative algorithm to solve such an inverse problem; (iii) the conversion of such an iterative algorithm onto a network where each algorithm iteration corresponds to a network layer; and (iv) the optimization of the resulting network parameters using back-propagation based on a dataset. These methods have been applied to a wide range of signal and image processing problems such as compressive sensing[19], deblurring[20], super-resolution[22, 23], and many more, where it has been shown that they can deliver higher performance while offering interpertability in comparison to purely data-driven techniques[30].

An increasing challenge arising in inverse problems affecting the performance of these various approaches relates to the presence of mismatch between the actual model and the postulated one due to erroneous modelling of the forward model[24, 25] or other anomalies. For example, in applications such as SAR imaging such a mismatch may arise from the motion error of the aircraft due to air disturbance. Likewise, in applications such as MRI imaging mismatch may arise from unexpected movement or jitter of the target object.

In scenarios where the mismatch between the true model and the postulated one is fixed, it has been shown that model-based approaches to inverse problems can suffer considerable degradation[25]; in contrast, data-driven techniques (including model aware data driven ones such as learnt iterative soft threshold algorithm (LISTA)[17]) are less affected because the inverse mapping corresponding to the actual model can still be learnt given enough training data. However, when the mismatch between the true model and the postulated one is not fixed but varies potentially across measurements over time – such as in the aforementioned SAR and MRI imaging settings – the performance of data-driven techniques, including model-aware ones, can be severely compromised. In fact, it has been shown that, in the presence of mismatch, the performance of model-based approaches to inverse problems may be significantly better than existing deep learning based methods[30].

Therefore, a number of approaches have recently emerged to successfully tackle model mismatch in inverse problems. For example, [26] proposed a total-least squares recovery algorithm allowing the reconstruction of a high-dimensional vector from low-dimensional noisy measurements in the presence of mismatch in the forward measurement operator. In [27], the authors proposed a matrix-uncertainty generalized approximate message passing algorithm in a Bayesian framework with informative priors allowing the reconstruction of high-dimensional vectors with random mismatch. There also exists various approaches to deal with model mismatch in different practical applications. The authors in [28] formulated a dynamic magnetic resonance imaging problem as a blind compressive sensing problem, wherein the forward-model is assumed to be integrated with unknown mismatch. In SAR applications, an alternating minimization strategy was proposed in [29] to simultaneously estimate the model mismatch and the sparse coefficients from the SAR data.

However, to the best of our knowledge, there has been less progress in the development of data-driven approaches, including deep learning ones, to inverse problems that are robust to model mismatch. There are some works relating to the design of deep neural networks robust to adversarial attacks[31], which are small imperceptible perturbations to the input data that seriously degrade the performance of an otherwise high-performing data-driven approaches. These include techniques based on adversarial training[32], data pre-processing[33] and robust neural network design[34, 35]. However, robustness against model mismatches has not been addressed.

In this paper, we propose a new approach to design robust neural network architectures, leveraging unfolding techniques. This allows to solve the linear inverse problem subject to uncertainty in the measurement operator. In particular, by building upon a total least squares formulation of a inverse problem where one wishes to recover a high-dimensional vector from low-dimensional noisy measurements subject to model mismatch, we make the following contributions:

  • We first develop a (robust) iterative soft threshold algorithm (ISTA) based solver to the underlying total least squares problem using conventional proximal gradient.

  • We then map this robust ISTA solver onto a Robust lEarned Shrinkage Thresholding (REST) network using algorithm unfolding techniques whereby each robust ISTA iteration corresponds to a REST network layer.

  • We demonstrate via experiments that REST can outperform various model-based and data-driven algorithms in compressive sensing and radar imaging problems.

The remainder of the paper is organized as follows: Section II formulates the linear inverse problem in the presence of mismatch in measurement operator. Section III describes the design of our proposed REST network. Sections IV and V present experimental results in compressive sensing and radar imaging tasks, respectively. Finally, in section VI, we draw conclusions. The Appendix A presents experimental results in compressive sensing task to compare the performance of the proposed REST network with shared and unshared parameters in each layer. The Appendix B presents some alternative strategies to unfold REST network along with their performance in a simple compressive sensing task. The Appendix C presents experimental results in compressive sensing task to support our intuitions on the comparison between REST and LISTA.

II Problem formulation

Consider a conventional linear inverse problem given by:

y=Ax+e,\displaystyle y=Ax+e, (1)

where yM×1y\in{\mathbb{R}}^{M\times 1} is an observation vector, xN×1x\in{\mathbb{R}}^{N\times 1} is the vector of interest, eM×1e\in{\mathbb{R}}^{M\times 1} is measurement noise, and AM×NA\in{\mathbb{R}}^{M\times N} models the inverse problem forward operator where N>MN>M. The goal is to reconstruct xx from the observation yy given a measurement matrix AA. These problems arise in various domains in signal and image processing, such as medical imaging systems[37], high-speed videos[38], single-pixel cameras[39], and radar imaging[40].

It is not generally possible to recover xx from yy when N>MN>M unless one makes additional assumptions about the structure of the vector of interest. In particular, by postulating that xx is sparse, a popular approach to recover xx from yy involves using the least-absolute shrinkage and selection operator (Lasso)[41], which is a solution to

minx\displaystyle\min_{x} yAx2+λx1,\displaystyle\quad\|y-Ax\|_{2}+\lambda\|x\|_{1}, (2)

where 2\|\cdot\|_{2} is the l2l_{2} norm and 1\|\cdot\|_{1} denotes the l1l_{1}-norm. This optimization problem can be solved using the well-known iterative soft thresholding algorithm (ISTA)[42] or ADMM[43].

Alternatively, one can adopt algorithm unfolding or unrolling techniques[20] to map such solvers onto a neural network architecture, whose parameters can then be further tuned using gradient descent or some variant based on the availability of a series of examples {(xi,yi}i=1n\{(x_{i},y_{i}\}_{i=1}^{n}. Networks derived from ISTA or ADMM known as LISTA[17] and ADMM-Net[19] respectively, have been shown to perform much better than purely learnt networks (e.g.[16]) or ISTA[42] and ADMM[43].

Consider now a more challenging scenario where the observation vector yM×1y\in{\mathbb{R}}^{M\times 1} is related to xN×1x\in{\mathbb{R}}^{N\times 1} as:

y=(A+E)x+e.\displaystyle y=(A+E)x+e. (3)

The model in (1) differs from the model in (3) in that the forward operator AA - which is assumed to be known - is now contaminated by an error matrix EE, which is unknown. Therefore, one now needs to recover the vector of interest in the presence of model mismatch EE.

To recover xx from yy in this case, one may consider a robust version of LASSO, or an l1l_{1} regularized version of total least squares[26] as follows:

minx,E\displaystyle\min_{x,E} [Ee]2+λx1,\displaystyle\quad\|[E\quad e]\|_{2}+\lambda\|x\|_{1},
s.t.\displaystyle{\rm s.t.} y=(A+E)x+e.\displaystyle\quad y=(A+E)x+e. (4)

Our goal is to build upon this formulation in order to design a data-driven approach using algorithm unfolding or unrolling techniques to recover xx reliably from yy in the presence of model mismatch. Model mismatch can be divided into two categories according to whether the mismatch is varying or fixed for different samples:

  1. 1.

    sample-wise fixed mismatch: In this scenario, the model mismatch is fixed for different data samples. One can then reliably recover xx from yy using data driven approaches such as LISTA. The reason is that the inverse mapping corresponding to the actual model can be learnt given enough training data as varying samples share the same model mismatch.

  2. 2.

    sample-wise varying mismatch: In this scenario, the model mismatch is changing for different data samples, which is much more common in practical applications. It is very hard for existing data-driven approaches to deal with sample-wise varying model mismatch. There also appears to be less work on how to design data-driven approaches – especially deep learning ones – with sample-wise varying model mismatch.

Our goal is to design a data-driven approach using algorithm unfolding technique based on the formulation in (II) for sample-wise varying mismatch.

III Robust Learned Shrinkage-Thresholding (REST) Network

III-A Robust ISTA

As shown in [26], the problem (II) can be solved in two different ways:

  1. 1.

    The first approach involves simultaneous recovery of xx and EE given yy. This is possible by converting (II) into

    minx,E\displaystyle\min_{x,E} y(A+E)x22+EF2+λx1.\displaystyle\quad\|y-(A+E)x\|^{2}_{2}+\|E\|_{F}^{2}+\lambda\|x\|_{1}. (5)
  2. 2.

    By admitting the closed-form solution of EE, the second approach converts the original total least squares optimization problem in (II) into the following optimization problem only related to xx:

    minxyAx221+x22+λx1.\displaystyle\min_{x}\frac{\|y-Ax\|^{2}_{2}}{1+\|x\|^{2}_{2}}+\lambda\|x\|_{1}. (6)

In [26], the optimization problem in (6) is solved by a two iteration loop method, which comprises an outer iteration loop based on the bisection method[44], and an inner iteration loop that relies on a variant of branch-and-bound (BB) method[45]. Here, we propose a proximal gradient method in order to design an iterative algorithm to recover xx from yy, which will be utilized to unfold into a network. This proximal gradient method here can be seen as a robust version of the ISTA algorithm, because both ISTA and robust ISTA solve a l1l_{1} regularized version of a least square problem using proximal gradient methods. However, robust ISTA deals with the optimization problem appearing in (6), where model mismatch is present, whereas ISTA deals with the optimization problem in (2).

Taking the gradient on the first term in (6) and executing a proximal step on the second term results in

xk+1=𝒮μλ{xkμgk},\displaystyle x^{k+1}={\cal S}_{\mu\lambda}\left\{x^{k}-\mu g^{k}\right\}, (7)

where μ>0\mu>0 is a step size, xkx^{k} represents the estimation of xx in the kk-th iteration, and gkg^{k} denotes the gradient of the first term evaluated at xkx^{k}:

gk=2[(1+xk22)AT(Axky)yAxk22xk](1+xk22)2.\displaystyle g^{k}=\frac{2\left[(1+\|{x^{k}}\|_{2}^{2})A^{T}(Ax^{k}-y)-\|y-Ax^{k}\|_{2}^{2}x^{k}\right]}{(1+\|{x^{k}}\|_{2}^{2})^{2}}. (8)

The operator Sμλ()S_{\mu\lambda}(\cdot) is the soft thresholding operator, and is applied element-wise on its vector argument as

𝒮μλ{x}=sign(x)max(|x|μλ,0).\displaystyle{\cal S}_{\mu\lambda}\{x\}={\rm sign}(x)\cdot\max(|x|-\mu\lambda,0). (9)

We next adopt unfolding techniques in order to map the robust recovery algorithm in (7) onto a robust neural network that can be used to recover high-dimensional sparse vectors from low-dimensional noisy measurements in the presence of model mismatch.

III-B Robust Learned Shrinkage-Thresholding Network

We first use (8) in order to re-write (7) as follows:

xk+1=\displaystyle x^{k+1}=
𝒮μλ(xk+2μyAxk22(1+xk22)2xk2μATA1+xk22xk+2μAT1+xk22y).\displaystyle{\cal S}_{\mu\lambda}\left(x^{k}+\frac{2\mu\|y-Ax^{k}\|_{2}^{2}}{(1+\|{x^{k}}\|_{2}^{2})^{2}}x^{k}-\frac{2\mu A^{T}A}{1+\|{x^{k}}\|_{2}^{2}}x^{k}+\frac{2\mu A^{T}}{1+\|{x^{k}}\|_{2}^{2}}y\right). (10)

Replacing AA in the first term by A1A_{1}, ATAA^{T}A in the second term by A2A_{2}, and ATA^{T} in the third term by A3A_{3}, (III-B) becomes

xk+1=\displaystyle x^{k+1}=
𝒮μλ(xk+2μyA1xk22(1+xk22)2xk2μA21+xk22xk+2μA31+xk22y).\displaystyle{\cal S}_{\mu\lambda}\left(x^{k}+\frac{2\mu\|y-A_{1}x^{k}\|_{2}^{2}}{(1+\|{x^{k}}\|_{2}^{2})^{2}}x^{k}-\frac{2\mu A_{2}}{1+\|{x^{k}}\|_{2}^{2}}x^{k}+\frac{2\mu A_{3}}{1+\|{x^{k}}\|_{2}^{2}}y\right). (11)

We now map the iterative algorithm in (III-B) onto a neural network architecture using unfolding techniques. In particular, we firstly map each iteration of the algorithm onto a layer of the neural network. Such a layer produces an output xk+1x^{k+1} – which is input to the next layer – based on an input xkx^{k} that undergoes a series of transformation such as linear transformations, linear scalings, and soft-thresholding operations. We then stack various such layers in order to produce an overall neural network. Such a network consisting of KK layers corresponds to KK iterations of the original robust ISTA algorithm, it accepts as input an initialization x0x^{0}, and it delivers as output an approximate solution xKx^{K}. Finally, we use different learnable parameters per network layer corresponding to the original parameters. Concretely, we choose learnable parameters λ,μ,A1,A2\lambda,\mu,A_{1},A_{2} and A3A_{3}. Note that each layer in the proposed network share the same learnable parameters (i.e., λ,μ,A1,A2\lambda,\mu,A_{1},A_{2} and A3A_{3}) are set to be the same in each layer) to give the network more restrictions to promote a better performance. We compare the performance of shared and unshared learnable parameters in each layer of REST in Appendix A.

Refer to caption
Figure 1: REST Neural Network Architecture.

The architecture of our network is depicted in Fig. 1. We note that depending on how we define the learnable parameters associated with (III-B) and (III-B) we may have slightly different neural network architectures. We elaborate on these possibilities in Appendix B.

III-B1 Learning Algorithm

To optimize the learnable parameters we rely on a dataset containing a series of pairs (xl,ylx_{l},y_{l}), l=1,,Ll=1,...,L where xlx_{l} corresponds to the original sparse vector and yly_{l} corresponds to the observation vector derived from the linear model in (3) for some fixed forward operator AA and some unknown forward operator perturbation EE, that may vary from example to example. Therefore, by fixing AA across training examples and letting EE and nn vary across training examples, we learn a network that is able to recover xx from y=(A+E)x+ny=(A+E)x+n for any unknown model perturbation EE.

We rely on a standard cost function measuring the difference between the network prediction of the vector of interest and the ground truth vector of interest as follows:

minθ1Ll=1Lx^lxl22\displaystyle\min_{\theta}\frac{1}{L}\sum_{l=1}^{L}\|\hat{x}^{l}-x^{l}\|^{2}_{2} (12)

where x^l\hat{x}^{l} corresponds to the network output given network input yly^{l}, xlx^{l} corresponds to the ground truth and θ\theta aggregates the various learnable parameters. We then adopt standard gradient descent algorithms in order to learn the network parameters λ,μ,A1,A2\lambda,\mu,A_{1},A_{2} and A3A_{3}.

III-C REST vs LISTA

Refer to caption
Figure 2: LISTA Neural Network Architecture.

We show in sequel that REST can perform substantially better than other unfolding approaches such as LISTA in the presence of inverse problems subject to model uncertainty. It is therefore instructive to compare the REST network architecture to the LISTA architecture (cf. Fig. 1 vs Fig. 2). The main difference between a layer in LISTA and in REST lies in two additional processing operations: 1) one corresponds to the second term in (III-B) and 2) the other corresponds to the normalization by 1+x221+\|x\|^{2}_{2}. In fact, without these operations, the REST architecture immediately reduces to a network very similar to LISTA. These additional operations play a critical role in mitigation of sample-wise model mismatch. The first operation enlarges the number of parameters, allowing for a much better fit. In REST, we have 2M×N+N×N+22M\times N+N\times N+2 learnable parameters in each layer and comparatively, we have M×N+N×N+1M\times N+N\times N+1 learnable parameters for LISTA. The second operation normalizes the output of each layer, promoting a more robust solution.

We perform various experiments in Appendix C to support our intuitions on the comparison between REST and LISTA.

IV Numerical Results

We now conduct a series of experiments on a simple compressive sensing task where one wishes to recover a vector of interest from a noisy linear measurements, in the presence of model mismatch. We compare the performance of the proposed REST network to well-known LISTA and ADMM-Net as well as results for Basis Pursuit (BP)[47], ISTA[42], robust ISTA[26], and MU-GAMP[27] algorithms.

IV-A Experimental setup

Our experimental set-up involves the generation of synthetic data as follows: We generate 1100 different sample pairs (x,y)(x,y) where the matrix is AA remains fixed, the matrix EE varies from sample to sample, and the noise vector ee and target vector xx also vary from sample to sample. Up to 1000 of these sample pairs is used for training purposes for the learning-based approaches, i.e., REST, LISTA and ADMM-Net, and the remaining 100 samples are used for testing purposes (i.e. evaluation of the performance of the approaches) for all the different methods. The measurement vector y13y\in\mathbb{R}^{13} is generated per (3), the target vector x250x\in\mathbb{R}^{250} is sparse with four randomly chosen non-zero elements taking values uniformly in the interval (0,1], and the noise vector e130e\in\mathbb{R}^{130} has i.i.d. Gaussian entries with zero mean and variance σ2=0.03\sigma^{2}=0.03. The matrix A130×250A\in\mathbb{R}^{130\times 250} has i.i.d. Gaussian entries with zero mean and variance such that AF=10\|A\|_{F}=10. The matrix E130×250E\in\mathbb{R}^{130\times 250} also has i.i.d. Gaussian entries with zero mean; furthermore, for samples in the training set we generate perturbation matrices EE such that EFr\|E\|_{F}\leq r and for samples in the testing set we generate perturbation matrices EE such that EFr\|E\|_{F}\leq r^{\prime}. This allows us to compare the performance of different learning-based approaches in scenarios where (a) there is no model mismatch in the training samples, in the testing samples, or across training and testing samples (r=r=0r=r^{\prime}=0) and (b) there is model mismatch in training samples, testing samples, or across training and testing samples (r0r\neq 0 or r0r^{\prime}\neq 0).

We use the reconstruction mean squared error (MSE) to evaluate the performance of the different approaches.

MSE=l=1L1LNxlx^l22,\displaystyle{\rm MSE}=\sum_{l=1}^{L}\frac{1}{LN}\|x^{l}-\hat{x}^{l}\|^{2}_{2}, (13)

where x^l\hat{x}^{l} corresponds to the network output given network input yly^{l}, xlx^{l} corresponds to the ground truth.

We consider the various networks listed in Table I. The learnable parameters of these networks in each layer are tied with shared parameters in each layer. Note that REST6, ADMM-Net8 and LISTA8 have similar parameter numbers, i.e., 765,012 parameters for REST6, 760,008 parameters for LISTA8 and 760,016 parameters for ADMM-Net8. However, the numbers of learnable parameters are different, i.e., 127,502 learnable parameters for REST6, 95,001 learnable parameters for LISTA8 and 95,002 learnable parameters for ADMM-Net8 as each layer share the same parameters.

TABLE I: Networks used in Experiments.
REST6 REST with 6 layers
LISTA6 LISTA with 6 layers
LISTA8 LISTA with 6 layers
ADMM-Net6 ADMM-Net with 6 layers
ADMM-Net8 ADMM-Net with 8 layers

We then train the various networks by using stochastic gradient descent (SGD) with learning rate lr=104l_{r}=10^{-4} and batch size to be 32. We use 2000 epochs to train the networks.

IV-B Experimental Results

Fig. 3 depicts the training and testing loss in terms of the number of training epochs for the different networks shown in Table I for r=r=5r=r^{\prime}=5 with 1000 training samples and 100 testing samples. We observe that LISTA and ADMM-Net converge faster than REST. However, REST fits the data much better than LISTA and ADMM-Net in the presence of model mismatch. In addition, the testing error in REST is lower than LISTA and ADMM-Net, and the generalization error - corresponding to the difference between testing and training errors - of REST is smaller than that of LISTA and ADMM-Net as the number of epochs increase. Notably, these trends also apply to other values of rr and rr^{\prime}.

Fig. 4 depicts training and testing loss vs number of training samples of the different networks shown in Table I for r=r=5r=r^{\prime}=5 with 100 testing samples and 2000 epochs. We note again that training error is much lower for REST than LISTA and ADMM-Net, especially for a larger number of training points. This indicates that REST fits the training data much better than LISTA and ADMM-Net. Likewise, the testing error is also much lower for REST in comparison with LISTA and ADMM-Net. Fig. 4 also shows that in the presence of model mismatch, the generalization error is smaller for REST than LISTA and ADMM-Net. Similar trends also apply for different values of rr and rr^{\prime}.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 3: Training and testing loss vs number of epochs for r=r=5r=r^{\prime}=5 with 1000 training samples and 100 testing samples. (a) REST6. (b). LISTA6. (c)LISTA8. (d). ADMM-Net6. (e)ADMM-Net8.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 4: Training and testing loss vs training set size with 100 testing samples and 2000 epochs r=r=5r=r^{\prime}=5 with 2000 epochs. (a) REST6. (b). LISTA6. (c)LISTA8. (d). ADMM-Net6. (e)ADMM-Net8.

We next evaluate the performance of different approaches as a function of the degree of model mismatch present in training data (rr) and testing data (rr^{\prime}). Table II depicts training errors (10log1010\log_{10}MSE) of REST, LISTA and ADMM-Net for different mismatch errors existing in the training data with 1000 training samples and 2000 epochs. It can be observed that when the r=0r=0 (i.e. when the forward model is fixed so that each training sample sees the same forward model), LISTA and ADMM-Net have smaller training errors than REST. This is because LISTA and ADMM-Net apply to inverse problems without any model mismatch. However, when the r2r\geq 2, LISTA and ADMM-Net do not fit well to the training data whereas REST does much better.

TABLE II: Training loss (10log1010\log_{10}MSE) of different networks with different rr with 1000 training samples and 2000 epochs.
r=0r=0 r=2r=2 r=4r=4 r=6r=6 r=8r=8 r=10r=10 r=12r=12
REST6 -2.61 -1.98 -1.67 -1.28 -1.06 -0.94 -0.81
LISTA6 -3.70 -1.14 -0.97 -0.61 -0.34 0.29 0.45
LISTA8 -3.74 -1.15 -1.02 -0.66 -0.36 0.28 0.44
ADMM-Net6 -3.96 -1.57 -1.01 -0.70 -0.48 0.12 0.35
ADMM-Net8 -3.98 -1.57 -1.05 -0.73 -0.49 0.11 0.33

Table III depicts the testing errors (10log1010\log_{10}MSE) of different networks with different rr and rr^{\prime} with 1000 training samples, 100 testing samples and 2000 epochs. We now comment on the performance under different mismatch regimes:

  • Scenario r=r=0r=r^{\prime}=0: This scenario is such that the training samples and the testing samples are subject to the same forward operator. It is clear that LISTA and ADMM-Net have better performance than REST because LISTA and ADMM-Net are specifically designed for inverse problems without any mismatch.

  • Scenario r0r\neq 0 or r0r^{\prime}\neq 0: This scenario is such that the training samples and/or the testing samples are subject to different forward operators. We can observe that – for a certain level of mismatch in the testing samples r>0r^{\prime}>0 – the higher the amount of mismatch in the training set, the higher the performance of the various unrolling approaches, i.e. REST, LISTA, and ADMM-Net. This is due to the fact that optimizing any of the networks with training data that is subject to different models inherently leads to a more robust network (i.e. this is an instance of robust training). We also observe that for a certain level of mismatch in the training samples r>0r>0 – the higher the amount of model mismatch in the testing samples the lower the performance of the various networks. Importantly, both in scenarios where r>rr>r^{\prime}, r<rr<r^{\prime}, or r=r0r=r^{\prime}\neq 0, the proposed REST network outperforms any of the competing networks.

TABLE III: Testing loss (10log1010\log_{10}MSE) for different networks with different rr and rr^{\prime} with 1000 training samples, 100 testing samples and 2000 epochs.
Methods r=0r=0 r=4r=4 r=8r=8
r=0r^{\prime}=0 r=4r^{\prime}=4 r=8r^{\prime}=8 r=12r^{\prime}=12 r=0r^{\prime}=0 r=4r^{\prime}=4 r=8r^{\prime}=8 r=12r^{\prime}=12 r=0r^{\prime}=0 r=4r^{\prime}=4 r=8r^{\prime}=8 r=12r^{\prime}=12
REST6 -2.57 -1.58 -0.99 -0.74 -2.40 -1.72 -1.22 -1.01 -2.05 -1.87 -1.52 -1.25
LISTA6 -3.49 -0.62 0.48 0.54 -2.57 -0.74 -0.26 0.09 -1.60 -0.78 -0.47 -0.22
LISTA8 -3.49 -0.65 0.47 0.52 -2.56 -0.79 -0.27 0.07 -1.62 -0.81 -0.53 -0.27
ADMM-Net6 -3.77 -0.53 0.62 0.81 -2.91 -1.03 -0.48 -0.13 -1.77 -1.01 -0.66 -0.40
ADMM-Net8 -3.78 -0.54 0.61 0.80 -2.90 -1.05 -0.51 -0.16 -1.77 -1.01 -0.69 -0.41

Table IV also depicts the testing errors (10log1010\log_{10}MSE) of different model-based approaches.

TABLE IV: Errors (10log1010\log_{10}MSE) of testing data for different model-based approaches with different rr^{\prime}..
r=0r^{\prime}=0 r=4r^{\prime}=4 r=8r^{\prime}=8 r=12r^{\prime}=12
BP -3.07 -1.26 0.45 0.91
ISTA -2.82 -0.73 0.43 0.83
Robust ISTA -2.47 -1.22 -0.48 -0.17
MU-GAMP -2.39 -1.30 -0.99 -0.70

As expected, comparing Table III to Table IV, without any model mismatch in the testing data (r=0r^{\prime}=0), model-based approaches outperform REST (independently of the value of rr); conversely, with model mismatch in the testing data (r>0r^{\prime}>0), REST outperforms any of the model-based approaches, for any value of rr (the level of model mismatch in the training data).

Refer to caption
Figure 5: Performance of REST with different layer numbers with 1000 training samples, 100 testing samples and 2000 epochs

We also evaluate the performance of the proposed REST network as a function of the number of layers. We use the aforementioned default experiment setting except that we now vary the number of layers from 1 to 10. It is clear from Fig. 5 that the number of layers does not appear to affect significantly the training loss. In contrast, it is also clear that the number of layers can affect the testing loss. We attribute this observation to the fact that unfolding networks architecture comes with a strong inductive bias, so that training data can be well fitted with a small number of layers. However, when the number of network layers is small, the processing capacity of REST for complex data is poor. Although REST with a small number of layers can fit training data well, the performance is poor when it comes to the new data during testing procedure.

Finally, we evaluate the computational cost of the different approaches. The different algorithms were implemented on a work server with 4 NVIDIA Tesla V100 GPU. The running times (seconds) to train the networks REST6, LISTA6, LISTA8, ADMM-Net6 and ADMM-Net8 are 41.7s, 29.8s, 39.4s, 26.2s and 37.1s, respectively, while the running times of using the learnt networks to predict the results on testing data of REST6, LISTA6, LISTA8, ADMM-Net6 and ADMM-Net8 are 3.0×104\times 10^{-4}s, 7.8×105\times 10^{-5}s, 1.0×104\times 10^{-4}s, 7.9×105\times 10^{-5}s and 1.0×104\times 10^{-4}s, respectively. The running time of BP, ISTA, robust ISTA and MU-GAMP algorithms are 8.3×103\times 10^{-3}s, 8.9×103\times 10^{-3}s, 2.2×102\times 10^{-2}s and 3.7×102\times 10^{-2}s, respectively. Although the computational cost of the training phase is relatively high, optimized networks deliver estimation results much faster than the model-based approaches. Notably, training complexity of REST is not considerably higher than that of LISTA or ADMM-Net, with the slightly higher complexity offset by much better performance.

V REST in radar imaging

We finally apply the proposed REST network in a synthetic aperture radar (SAR) imaging application subject to model mismatch.

V-A Problem formulation

We consider a SAR imaging problem where the radar transmits linear frequency modulated (LFM) pulses at a constant rate. Let xmx_{m} and yny_{n} refer to the mmth range gate and nnth azimuth bin of the target SAR image ZZ, respectively, where m=1,2,,Mm=1,2,\cdots,M and n=1,2,,Nn=1,2,\cdots,N. MM and NN represent the sample numbers of range and azimuth direction, respectively. It is assumed that τ\tau and tt respectively represent the range and azimuth time, and i=1,2,,Ii=1,2,\cdots,I and q=1,2,,Qq=1,2,\cdots,Q index the range and azimuth time, where II and QQ indicate the sample number of range and azimuth time, respectively. Set s(τi,tq,xm,yn)s(\tau_{i},t_{q},x_{m},y_{n}) as the received signal of the qqth pulse in the iith sequence from the target located at (xm,yn)(x_{m},y_{n}), which can be written as follows:

s(τi,tq,xm,yn)=φ(τi,tq,xm,yn)Z(xm,yn),s(\tau_{i},t_{q},x_{m},y_{n})=\varphi(\tau_{i},t_{q},x_{m},y_{n}){{Z}}(x_{m},y_{n}), (14)

where Z(xm,yn){{Z}}(x_{m},y_{n}) denotes the reflector coefficient of the target located at (xm,yn)(x_{m},y_{n}), and

φ(τi,tq,xm,yn)=\displaystyle\varphi(\tau_{i},t_{q},x_{m},y_{n})= ωr(τi)ωa(tq)\displaystyle{\omega_{r}}\left({{\tau_{i}}}\right){\omega_{a}}\left({{t_{q}}}\right)
exp{jπkr(τiR(tq,xm,yn)c)2}\displaystyle\cdot\exp\left\{{-j\pi{k_{r}}{{\left({{\tau_{i}}-\frac{{{R}({t_{q}},x_{m},y_{n})}}{c}}\right)}^{2}}}\right\}
exp{j4πfcR(tq,xm,yn)c}.\displaystyle\cdot\exp\left\{{-j4\pi{f_{c}}\frac{{{R}({t_{q}},x_{m},y_{n})}}{c}}\right\}. (15)

In (V-A), j=1j=\sqrt{-1}, fcf_{c} is the carrier frequency, cc is the speed of light, krk_{r} indicates the frequency modulation (FM) rate associated with the transmitted LFM signal, ωr{\omega_{r}} and ωa{\omega_{a}} denote range and azimuth envelope corresponding to rectangular window functions given by

ωr(τi)={1,if|τi2R(tq,xm,yn)/c|<Tr2,0,otherwise,\displaystyle{\omega_{r}}(\tau_{i})=\left\{{\begin{array}[]{*{20}{c}}1,\quad{\rm if}\quad|\tau_{i}-{2{R}({t_{q}},x_{m},y_{n})}/{c}|<\frac{T_{r}}{2},\\ 0,\quad{\rm otherwise},\end{array}}\right. (18)

and

ωa(tq)={1,if|tqyn/v|<Ts2,0,otherwise.\displaystyle{\omega_{a}}(t_{q})=\left\{{\begin{array}[]{*{20}{c}}1,\quad{\rm if}\quad|t_{q}-{y_{n}}/{v}|<\frac{T_{s}}{2},\\ 0,\quad{\rm otherwise}.\end{array}}\right. (21)

where TrT_{r} and TsT_{s} denote the transmitted signal time width and synthetic aperture time, respectively, and R(tq,xm,yn){{R}({t_{q}},x_{m},y_{n})} represents the instantaneous slant range from image position (xm,yn)(x_{m},y_{n}) to the moving platform at azimuth time tqt_{q}

R(tq,xm,yn)=(yp+vtqyn)2+(xpxm)2+zp2,\displaystyle{R}({t_{q}},x_{m},y_{n})=\sqrt{(y_{p}+vt_{q}-y_{n})^{2}+(x_{p}-x_{m})^{2}+z_{p}^{2}}, (22)

where (xp,yp,zp)(x_{p},y_{p},z_{p}) denotes radar platform position at the time tq=0t_{q}=0, and the radar platform is assumed to move along yy axis with constant velocity vv. The corresponding SAR system configuration is illustrated in Fig. 6.

Refer to caption
Figure 6: Configuration of SAR system.

Assume the received signal at the qqth pulse of the iith sequence is represented by S(τi,tq){{S}}(\tau_{i},t_{q}), which can be written as follows:

S(τi,tq)\displaystyle{{S}}(\tau_{i},t_{q}) =mns(τi,tq,xm,yn)+Γ(τi,tq)\displaystyle=\sum\limits_{m}{\sum\limits_{n}{s(\tau_{i},t_{q},x_{m},y_{n})}}+\Gamma(\tau_{i},t_{q})
=mnφ(τi,tq,xm,yn)Z(xm,yn)+Γ(τi,tq),\displaystyle=\sum\limits_{m}{\sum\limits_{n}{\varphi(\tau_{i},t_{q},x_{m},y_{n})Z(x_{m},y_{n})}}+\Gamma(\tau_{i},t_{q}), (23)

where Γ(τi,tq)\Gamma(\tau_{i},t_{q}) denotes the additive noise at (τi,tq)(\tau_{i},t_{q}). The equation can be also rewritten as

y=Ax+e,\displaystyle y=Ax+e, (24)

where y=vec(S)y={\rm vec}({{S}}), x=vec(Z)x={\rm vec}({{Z}}), e=vec(Γ)e={\rm vec}(\Gamma), HH is

A=[φ1,1,1,1φ1,1,2,1φ1,1,M,Nφ2,1,1,1φ2,1,2,1φ2,1,M,NφI,Q,1,1φI,Q,2,1φI,Q,M,N],\displaystyle A=\left[{\begin{array}[]{*{20}{c}}\varphi_{1,1,1,1}&\varphi_{1,1,2,1}&\cdots&\varphi_{1,1,M,N}\\ \varphi_{2,1,1,1}&\varphi_{2,1,2,1}&\cdots&\varphi_{2,1,M,N}\\ \vdots&\vdots&\ddots&\vdots\\ \varphi_{I,Q,1,1}&\varphi_{I,Q,2,1}&\cdots&\varphi_{I,Q,M,N}\\ \end{array}}\right], (29)

and φi,q,m,n=φ(τi,tq,xm,yn)\varphi_{i,q,m,n}=\varphi(\tau_{i},t_{q},x_{m},y_{n}). Subsequently, the radar imaging problem can be cast onto the problem:

minx\displaystyle\min_{x} yAx2+λx1.\displaystyle\quad\|y-Ax\|_{2}+\lambda\|x\|_{1}. (30)

In the practical application of SAR imaging, because of influence the of platform vibration, wind field, and turbulence, the antenna phase center of the radar platform might significantly deviate from the ideal trajectory, thus leading to errors in terms of instantaneous slant range R(tq,xm,yn){{R}({t_{q}},x_{m},y_{n})}. Errors in R(tq,xm,yn){{R}({t_{q}},x_{m},y_{n})} introduce mismatch in the nominal measurement AA appearing in (24) and (30). In principle, it is possible to calculate errors of R(tq,xm,yn){{R}({t_{q}},x_{m},y_{n})} using motion measurement data recorded by an ancillary measurement instrument such as inertial navigation system (INS), inertial measurement units (IMUs), and global positioning system (GPS). However, these measuring instruments are not precise enough for the required accuracy, and in many cases, there are no usable motion measurement data.

Suppose the slant range is

R(tq,xm,yn)=\displaystyle{R}^{\prime}({t_{q}},x_{m},y_{n})=
(yp+vtqynΔyq)2+(xpxmΔxq)2+(zpΔzq)2,\displaystyle\sqrt{(y_{p}+vt_{q}-y_{n}-\Delta y_{q})^{2}+(x_{p}-x_{m}-\Delta x_{q})^{2}+(z_{p}-\Delta z_{q})^{2}}, (31)

where Δxq\Delta x_{q}, Δyq\Delta y_{q} and Δzq\Delta z_{q} denote the motion deviations from the nominal trajectory at azimuth time tqt_{q} with respect to xx, yy and zz directions, respectively. Replacing R(tq,xm,yn){R}({t_{q}},x_{m},y_{n}) in (V-A) and (29) by R(tq,xm,yn){R}^{\prime}({t_{q}},x_{m},y_{n}) leads to a measurement model associated with the operator A+EA+E instead of AA only, where EE captures the difference between R(tq,xm,yn){R}({t_{q}},x_{m},y_{n}) and R(tq,xm,yn){R^{\prime}}({t_{q}},x_{m},y_{n}). The original model in (24) then becomes

y=(A+E)x+e,\displaystyle y=(A+E)x+e, (32)

implying that we can leverage REST for SAR imaging applications.

Note, however, that in SAR imaging problem, the variables in (32) are complex rather than real. To use the proposed REST network, we decompose the variable into their real and imaginary parts.

V-B Experimental Results

We compare the performance of the proposed REST network to LISTA, ADMM-Net, BP, ISTA and robust ISTA, and MU-GAMP algorithms.

V-B1 Experimental setup

System and geometry parameters are listed in Table V.

TABLE V: Simulation parameters
𝐏𝐚𝐫𝐚𝐦𝐞𝐭𝐞𝐫\mathbf{Parameter} 𝐕𝐚𝐥𝐮𝐞\mathbf{Value}
fcf_{c}: Carrier frequency 10GHz10\mathrm{GHz}
BrB_{r}: Signal bandwidth 400MHz400\mathrm{MHz}
TrT_{r}: Signal time width 1.5μs1.5\mathrm{\mu s}
fsf_{s}: Fast time sampling rate 500MHz500\mathrm{MHz}
TsT_{s}: Synthetic aperture time 2s2\mathrm{s}
vv: Radar platform velocity 200m/s200m/s
PRFPRF: Pulse repetition frequency 800Hz800\mathrm{Hz}
xpx_{p}: Radar position in xx direction at t=0t=0 3km3\mathrm{km}
ypy_{p}: Radar position in yy direction at t=0t=0 0km0\mathrm{km}
zpz_{p}: Radar position in zz direction at t=0t=0 5km5\mathrm{km}

Here, we randomly generate Δxq\Delta x_{q}, Δyq\Delta y_{q} and Δzq\Delta z_{q} in (V-A) with the constraint Δxq[d,d]\Delta x_{q}\in[-d,d], Δyq[d,d]\Delta y_{q}\in[-d,d] and Δzq[d,d]\Delta z_{q}\in[-d,d]. Note that different values of dd correspond to different rr and rr^{\prime}.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 7: SAR images used in training process.

In the training process, 400 different SAR images are used as reflector coefficient matrix ZZ, which are obtained by dividing each image in Fig. 7 (a)-(d) into 100 small images. As for the testing data, real SAR images are used as reflector coefficient matrix ZZ by dividing the image in Fig. 8 into 256 small images. We set the size of single data as M=512M=512 and N=128N=128. Both real and imagery parts of noise vector η\eta have i.i.d. Gaussian entries with zero mean and variance σ2=0.01\sigma^{2}=0.01. Then, values of yy, AA, EE, xx and ee in radar imaging task can be generated for both training and testing data.

Refer to caption
Figure 8: SAR image used in testing process.
TABLE VI: Definitions of different networks used in Experiments.
REST4 REST with 4 layers
LISTA4 LISTA with 4 layers
LISTA7 LISTA with 7 layers
ADMM-Net4 ADMM-Net with 4 layers
ADMM-Net7 ADMM-Net with 7 layers

We consider the various networks listed in Table VI, wherein the learnable parameters of these networks in each layer are set to be tied. Note that REST4, ADMM-Net7 and LISTA7 have the similar parameter numbers, i.e., 1,572,872 parameters for REST4, 2,293,767 parameters for LISTA7 and 2,293,774 parameters for ADMM-Net7. However, the numbers of learnable parameters are different, i.e., 393,218 learnable parameters for REST4, 327,681 learnable parameters for LISTA7 and 327,682 learnable parameters for ADMM-Net7.

We set two different training approaches to train REST4, LISTA4, LISTA7, ADMM-Net4 and ADMM-Net7:

  • Training strategy 1: When generating training data, we set d=0d=0 (hence E=0E=0). In this case, there is no mismatch in the training data. We are interested in assessing how the performance of different designed networks compare in the presence of model mismatch at testing time without any form of training robustization.

  • Training strategy 2: When generating training data, we set d=0.16d=0.16, and generate 400 different mismatch matrices. Among them, we observe that the maximum value of EF\|E\|_{F} is 3.01, and therefore, r=3.01r=3.01. In this case, we model mismatch in both training and testing data in the sense that mismatch changes from sample to sample, so we are interested in assessing how the performance of different networks compare in the presence of model mismatch at testing time with a robust training approach.

The various networks are also trained by using SGD with learning rate lr=4×104l_{r}=4\times 10^{-4} and batch size to be 32. We use totally 2000 epochs to train the networks.

V-B2 Experimental Results

Refer to caption
Refer to caption
Figure 9: Training loss vs number of training epochs for different networks. (a). Training strategy 1. (b). Training strategy 2.

The evolution of the training loss in terms of the number of training epochs of different networks under training strategies 1 and 2 are shown in Fig. 9 (a) and (b), respectively. It can be observed that in both training strategies 1 and 2, LISTA and ADMM-Net converge faster. However, when the training data contains model mismatch in training strategy 2, REST fits the data much better than LISTA and ADMM-Net.

The imaging results of different networks under training strategy 1 and training strategy 2 are shown in Fig. 10 and 11, respectively. Note that in Fig. 10 and 11, in line with standard practice, we present the amplitude value of σ\sigma and we combined the small images of testing data into a whole large image in order to better display the results. Imaging results by model-based approaches, i.e., BP, ISTA, robust ISTA and MU-GAMP, are shown in Fig. 12. In Fig. 10, 11 and 12, columns 1 to 5 correspond to reconstructed results of d=0d=0 and r=0r^{\prime}=0, d=0.18d=0.18 and r=1.05r^{\prime}=1.05, d=0.31d=0.31 and r=2.04r^{\prime}=2.04, d=0.45d=0.45 and r=2.97r^{\prime}=2.97, d=0.54d=0.54 and r=4.01r^{\prime}=4.01 for the testing data, respectively. In Fig. 10 and 11, Rows 1 to 5 correspond to REST4, LISTA4, LISTA7, ADMM-Net4 and ADMM-Net7, respectively. In Fig. 12, rows 1 to 4 correspond to BP, ISTA, robust ISTA and MU-GAMP algorithms, respectively.

From Fig. 10, 11 and 12, we observe that for the learning-based approaches like REST, LISTA and ADMM-Net, the performance is better when the networks are trained in strategy 2. It shows that when mismatch is taken into consideration during training, the robustness of the networks to mismatch error is improved. In strategy 1, the reconstructed results of LISTA and ADMM-Net become seriously blurred when r=2.04r^{\prime}=2.04, while the results of REST are blurred when r=2.97r^{\prime}=2.97. This is due to the fact that the REST is especially designed for inverse problem with mismatch even when there is no attempt to robustify the network during training. In strategy 2, the reconstructed results of LISTA and ADMM-Net are seriously blurred when r=2.97r^{\prime}=2.97, while results of REST are blurred when r=4.01r^{\prime}=4.01. When r=2.97r^{\prime}=2.97 and r=4.01r^{\prime}=4.01, results of REST are much better than those of LISTA and ADMM-Net under the same training strategy. This shows that REST is also much more immune to model mismatch even when one attempts to robustify the various networks during training.

Robust ISTA and MU-GAMP have better performance than BP and ISTA. However, the proposed REST outperforms Robust ISTA and MU-GAMP especially in the cases r=2.97r^{\prime}=2.97 and r=4.01r^{\prime}=4.01.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 10: Imaging results of different networks under training strategy 1. Columns 1 to 5 correspond to reconstructed results of d=0d=0 and r=0r^{\prime}=0, d=0.18d=0.18 and r=1.05r’=1.05, d=0.31d=0.31 and r=2.04r‘=2.04, d=0.45d=0.45 and r=2.97r’=2.97, d=0.54d=0.54 and r=4.01r‘=4.01 for the testing data, respectively. Rows 1 to 5 correspond to REST4, LISTA4, LISTA7, ADMM-Net4 and ADMM-Net7, respectively.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 11: Imaging results of different networks under training strategy 2. Columns 1 to 5 correspond to reconstructed results of d=0d=0 and r=0r^{\prime}=0, d=0.18d=0.18 and r=1.05r’=1.05, d=0.31d=0.31 and r=2.04r‘=2.04, d=0.45d=0.45 and r=2.97r’=2.97, d=0.54d=0.54 and r=4.01r‘=4.01 for the testing data, respectively. Rows 1 to 5 correspond to REST4, LISTA7, LISTA7, ADMM-Net4 and ADMM-Net7, respectively.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 12: Imaging results of different model-based approaches. Columns 1 to 5 correspond to reconstructed results of d=0d=0 and r=0r^{\prime}=0, d=0.18d=0.18 and r=1.05r’=1.05, d=0.31d=0.31 and r=2.04r‘=2.04, d=0.45d=0.45 and r=2.97r’=2.97, d=0.54d=0.54 and r=4.01r‘=4.01 for the testing data, respectively. Rows 1 to 5 correspond to BP, ISTA, robust ISTA and MU-GAMP algorithms, respectively.

Next, the image qualities corresponding to different algorithms and networks are measured quantitatively, using MSE of target vector xx to evaluate the reconstruction performance.

TABLE VII: MSE of different approaches with different rr^{\prime}. (r=0r=0 for strategy T1 and r=3.01r=3.01 for strategy T2)
Methods MSE
r=0r^{\prime}=0 r=1.05r^{\prime}=1.05 r=2.04r^{\prime}=2.04 r=2.97r^{\prime}=2.97 r=4.01r^{\prime}=4.01
REST4(T1) 0.0054 0.0077 0.0096 0.046 0.10
REST4(T2) 0.0077 0.0084 0.0097 0.013 0.074
LISTA4(T1) 0.0011 0.012 0.096 0.17 0.29
LISTA4(T2) 0.0069 0.0086 0.0097 0.059 0.19
LISTA7(T1) 0.0010 0.017 0.087 0.16 0.24
LISTA7(T2) 0.0052 0.0081 0.0094 0.051 0.16
ADMM-Net4(T1) 0.00097 0.0096 0.088 0.16 0.24
ADMM-Net4(T2) 0.0074 0.082 0.0096 0.056 0.18
ADMM-Net7(T1) 0.00095 0.0097 0.087 0.15 0.22
ADMM-Net7(T2) 0.0072 0.0080 0.0093 0.050 0.16
BP 0.0025 0.058 0.092 0.29 0.48
ISTA 0.0022 0.063 0.095 0.25 0.44
Robust ISTA 0.0069 0.0097 0.032 0.097 0.21
MU-GAMP 0.0077 0.0092 0.029 0.090 0.19

The performance of different approaches is listed in Table VII, where T1 and T2 in the brackets represent that the networks are trained using strategy 1 and strategy 2, respectively. The observations obtained from Fig. 10, 11 and 12 could be verified by Table VII as well. In addition, we also note that:

  • LISTA7 and ADMM-Net7 have slightly better performance than LISTA4 and ADMM-Net4 especially when the mismatch in testing data is large. This indicates that increasing the number of layers in LISTA and ADMM-Net networks can improve the robustness, however, the improvement is very limited.

  • The proposed REST network is much more robust to the mismatch error than the existing learning- and model-based approaches. Even REST trained in training strategy 1 has a better performance than LISTA and ADMM-Net trained in training strategy 2 and other model-based approaches.

VI Conclusion

Deep learning has achieved significant successes in various signal and image processing tasks, but it is also vulnerable to various perturbations. By adopting algorithm unrolling techniques and proposing a REST network, we show that it is possible to design neural network architectures that can reliably recover high-dimensional sparse vectors from low-dimensional noisy linear measurements in the presence of challenging sample-wise model mismatch. We also apply the proposed REST network to solve the synthetic aperture radar imaging problem, wherein model mismatch occurs due to motion uncertainty of the moving radar platform. Experiments on both compressive sensing and synthetic aperture radar imaging tasks show the proposed REST network outperforms state-of-the-art learning-based and model-based approaches in view of various additional operations that seem to endow the architecture with inherent robustness to mismatches.

VII Appendix A

This appendix briefly introduces the performance of the proposed REST network with shared or unshared weights in each layer. We run various experiments (in line with the experimental set-up reported in Section IV-A). In Table VIII, we compare the testing loss (10log1010\log_{10}MSE) of shared and unshared parameters of REST with different rr and rr^{\prime} with 1000 training samples, 100 testing samples and 2000 epochs, where REST-I and REST-II denote the shared and unshared parameters of REST, respectively. It can be observed from Table VIII that the networks with shared and unshared parameters have very similar performance. When the mismatch in testing data is large, i.e., r=8r^{\prime}=8 and r=12r^{\prime}=12, the shared parameters network has slightly better performance.

TABLE VIII: Testing loss (10log1010\log_{10}MSE) of shared and unshared parameters of REST with different rr and rr^{\prime} with 1000 training samples, 100 testing samples and 2000 epochs.
Methods r=0r=0 r=4r=4 r=8r=8
r=0r^{\prime}=0 r=4r^{\prime}=4 r=8r^{\prime}=8 r=12r^{\prime}=12 r=0r^{\prime}=0 r=4r^{\prime}=4 r=8r^{\prime}=8 r=12r^{\prime}=12 r=0r^{\prime}=0 r=4r^{\prime}=4 r=8r^{\prime}=8 r=12r^{\prime}=12
REST-I -2.57 -1.58 -0.99 -0.74 -2.40 -1.72 -1.22 -1.01 -2.05 -1.87 -1.52 -1.25
REST-II -2.62 -1.60 -0.99 -0.78 -2.41 -1.70 -1.23 -1.01 -2.01 -1.86 -1.51 -1.23

VIII Appendix B

This appendix briefly overviews other options to turn the robust ISTA algorithm appearing in (III-B) onto other neural network structures, depending on how one selects the learnable parameters.

Option 1: This option is based on re-writing (III-B) as follows:

xk+1=𝒮μλ(A1xk+A2y),\displaystyle x^{k+1}={\cal S}_{\mu\lambda}(A_{1}x^{k}+A_{2}y), (33)

where A1=I+2μyAxk22I(1+xk22)22μATA1+xk22A_{1}=I+\frac{2\mu\|y-Ax^{k}\|_{2}^{2}I}{(1+\|{x^{k}}\|_{2}^{2})^{2}}-\frac{2\mu A^{T}A}{1+\|{x^{k}}\|_{2}^{2}} and A2=2μAT1+xk22A_{2}=\frac{2\mu A^{T}}{1+\|{x^{k}}\|_{2}^{2}}. We then set μλ\mu\lambda, A1A_{1} and A2A_{2} as learnable parameters. Clearly, this leads to an unfolded network akin to LISTA, so this unfolding approach does not carry any advantage.

Option 2: This option is based on re-writing (III-B) as follows:

xk+1=𝒮μλ(xk\displaystyle x^{k+1}={\cal S}_{\mu\lambda}(x^{k} +2μ1(1+xk22)2xk\displaystyle+\frac{2\mu_{1}}{(1+\|{x^{k}}\|_{2}^{2})^{2}}x^{k}
2μA11+xk22xk+2μA21+xk22y)\displaystyle-\frac{2\mu A_{1}}{1+\|{x^{k}}\|_{2}^{2}}x^{k}+\frac{2\mu A_{2}}{1+\|{x^{k}}\|_{2}^{2}}y) (34)

where μ1=μyAxk22\mu_{1}=\mu\|y-Ax^{k}\|_{2}^{2} and we have replaced ATAA^{T}A in the third term by A1A_{1}, and ATA^{T} in the fourth term by A2A_{2}. Here, we set μλ\mu\lambda, μ1\mu_{1}, A1A_{1} and A2A_{2} as learnable parameters.

Option 3: This option is based on re-writing (III-B) as follows:

xk+1=𝒮μλ(xk\displaystyle x^{k+1}={\cal S}_{\mu\lambda}(x^{k} +2μyA1xk22(1+xk22)2xk\displaystyle+\frac{2\mu\|y-A_{1}x^{k}\|_{2}^{2}}{(1+\|{x^{k}}\|_{2}^{2})^{2}}x^{k}
2μA21+xk22xk+2μA31+xk22y)\displaystyle-\frac{2\mu A_{2}}{1+\|{x^{k}}\|_{2}^{2}}x^{k}+\frac{2\mu A_{3}}{1+\|{x^{k}}\|_{2}^{2}}y) (35)

where we have replaced AA in the second term by A1A_{1}, ATAA^{T}A in the third term by A2A_{2}, and ATA^{T} in the fourth term by A3A_{3}. Here, we set λ\lambda, μ\mu, A1A_{1}, A2A_{2} and A3A_{3} as learnable parameters. The option 3 leads to a network structure akin to the one deriving from (III-B) and (III-B).

Option 4: Finally, one can also re-write (III-B) as follows:

xk+1=𝒮μλ(xk\displaystyle x^{k+1}={\cal S}_{\mu\lambda}(x^{k} +2μyTyxk(1+xk22)2\displaystyle+\frac{2\mu y^{T}yx^{k}}{(1+\|{x^{k}}\|_{2}^{2})^{2}}
4μyTA1xkxk(1+xk22)2+2μxkTA2xkxk(1+xk22)2\displaystyle-\frac{4\mu y^{T}A_{1}x^{k}x^{k}}{(1+\|{x^{k}}\|_{2}^{2})^{2}}+\frac{2\mu x^{kT}A_{2}x^{k}x^{k}}{(1+\|{x^{k}}\|_{2}^{2})^{2}}
2μA2xk1+xk22+2μA3y1+xk22)\displaystyle-\frac{2\mu A_{2}x^{k}}{1+\|{x^{k}}\|_{2}^{2}}+\frac{2\mu A_{3}y}{1+\|{x^{k}}\|_{2}^{2}}) (36)

where we have replaced the matrix AA in the third term by A1A_{1}, ATAA^{T}A in the fourth and fifth terms by A2A_{2}, and ATA^{T} in the sixth term by A3A_{3}. Here, we set λ\lambda, μ\mu, A1A_{1}, A2A_{2} and A3A_{3} as learnable parameters.

We run various experiments (in line with the experimental set-up reported in Section IV-A) in order to test the performance of the different unfolding options. In Table X, we compare the testing loss for different unfolding strategies of REST, where REST2-6, REST2-8, REST3-6 and REST4-6 represent the different unfolding strategies shown in subsection III-B. The definitions are shown in Table IX. Note that REST2-8 has similar parameter number as those of REST3-6 and REST4-6, i.e., 7616 parameters for REST2-8 and 7662 parameters for REST3-6 and REST4-6. It could be observed from Table X that the networks unfolded by the third and fourth schemes have similar performance, while the performance of the network by the second unfolding strategy is slightly worse. This motivated us to adopt the unfolding approach deriving from (III-B) and (III-B).

TABLE IX: Different unfolding strategies for REST.
REST2-6 REST unfolded by the 2nd strategy with 6 layers
REST2-8 REST unfolded by the 2nd strategy with 8 layers
REST3-6 REST unfolded by the 3rd strategy with 6 layers
REST4-6 REST unfolded by the 4th strategy with 6 layers
TABLE X: Testing loss (10log1010\log_{10}MSE) of different unfolding strategies of REST with different rr and rr^{\prime} with 1000 training samples, 100 testing samples and 2000 epochs.
Methods r=0r=0 r=4r=4 r=8r=8
r=0r^{\prime}=0 r=4r^{\prime}=4 r=8r^{\prime}=8 r=12r^{\prime}=12 r=0r^{\prime}=0 r=4r^{\prime}=4 r=8r^{\prime}=8 r=12r^{\prime}=12 r=0r^{\prime}=0 r=4r^{\prime}=4 r=8r^{\prime}=8 r=12r^{\prime}=12
REST2-6 -2.92 -1.55 -0.90 -0.67 -2.76 -1.70 -1.11 -0.94 -2.69 -1.84 -1.23 -1.01
REST2-8 -2.94 -1.57 -0.94 -0.69 -2.76 -1.71 -1.14 -0.97 -2.73 -1.86 -1.24 -1.06
REST3-6 -2.57 -1.58 -0.99 -0.74 -2.40 -1.72 -1.22 -1.01 -2.05 -1.87 -1.52 -1.25
REST4-6 -2.59 -1.57 -0.99 -0.73 -2.41 -1.72 -1.22 -1.02 -2.05 -1.86 -1.52 -1.26

IX Appendix C

In this appendix, we run various experiments (in line with the experimental set-up reported in Section IV-A) to support our intuitions on the comparison between REST and LISTA. As mentioned in Section III-C, the main difference between a layer in LISTA and in REST lies in two additional processing operations: 1) one corresponds to the second term in (III-B) and 2) the other corresponds to the normalization operation by 1+x221+\|x\|^{2}_{2}.

Here, we design two networks to be used for comparison with LISTA and REST. The first one is to delete the second term in (III-B) leading to

xk+1=𝒮μλ(xk2μA21+xk22xk+2μA31+xk22y).\displaystyle x^{k+1}={\cal S}_{\mu\lambda}\left(x^{k}-\frac{2\mu A_{2}}{1+\|{x^{k}}\|_{2}^{2}}x^{k}+\frac{2\mu A_{3}}{1+\|{x^{k}}\|_{2}^{2}}y\right). (37)

Compared with LISTA, this network has the same parameter number with an additional normalization by 1+x221+\|x\|^{2}_{2}.

The second network reduces the normalization in (III-B) resulting in

xk+1=𝒮μλ(xk+2μyA1xk22xk2μA2xk+2μA3y).\displaystyle x^{k+1}={\cal S}_{\mu\lambda}\left(x^{k}+{2\mu\|y-A_{1}x^{k}\|_{2}^{2}}x^{k}-{2\mu A_{2}}x^{k}+{2\mu A_{3}}y\right). (38)

Compared with LISTA, the main difference is the second term in (38), which enlarges the number of parameters.

TABLE XI: Testing loss (10log1010\log_{10}MSE) of different networks with different rr and rr^{\prime} with 1000 training samples, 100 testing samples and 2000 epochs.
Methods r=0r=0 r=4r=4 r=8r=8
r=0r^{\prime}=0 r=4r^{\prime}=4 r=8r^{\prime}=8 r=12r^{\prime}=12 r=0r^{\prime}=0 r=4r^{\prime}=4 r=8r^{\prime}=8 r=12r^{\prime}=12 r=0r^{\prime}=0 r=4r^{\prime}=4 r=8r^{\prime}=8 r=12r^{\prime}=12
REST6 -2.57 -1.58 -0.99 -0.74 -2.40 -1.72 -1.22 -1.01 -2.05 -1.87 -1.52 -1.25
NI6 -2.77 -1.62 -0.69 -0.13 -2.44 -1.70 -0.87 -0.31 -1.88 -1.44 -1.92 -0.61
NI8 -2.78 -1.61 -0.70 -0.13 -2.43 -1.71 -0.87 -0.32 -1.71 -1.45 -1.90 -0.60
NII6 -3.50 -1.77 -0.73 -0.22 -2.78 -1.32 -1.08 -0.47 -1.97 -1.58 -0.97 -0.66
LISTA6 -3.49 -0.62 0.48 0.54 -2.57 -0.74 -0.26 0.09 -1.60 -0.78 -0.47 -0.22
LISTA8 -3.49 -0.65 0.47 0.52 -2.56 -0.79 -0.27 0.07 -1.62 -0.81 -0.53 -0.27

In Table XI, we compare the testing loss (10log1010\log_{10}MSE) of different networks mentioned above with different rr and rr^{\prime} with 1000 training samples, 100 testing samples and 2000 epochs, where REST6, LISTA 6 and LISTA8 are defined in Table I and NI6, NI8 and NII8 are the network in (37) with 6 layers, network in (37) with 8 layers and the network in (38) with 6 layers, respectively. REST6, NI8 and LISTA8 have similar parameter numbers, i.e., 765,012 parameters for REST6, 760,008 parameters for LISTA8 and NI8. However, the numbers of learnable parameters are different, i.e., 127,502 learnable parameters for REST6, 95,001 learnable parameters for LISTA8 and NI8. It could be observed from Table XI that both NI and NII have smaller testing errors than LISTA and larger testing errors than REST when there is model mismatch in testing data, which indicates that both two additional processing operations promote the robustness of the network.

References

  • [1] E. J. Candes, J. Romberg, and T. Tao, “Near-optimal signal recovery from random projections: Universal encoding strategies?,” IEEE Trans. Inf. Theory, vol. 52, no. 12, pp. 5406–5425, 2006.
  • [2] D. L. Donoho, “Compressed sensing,” IEEE Trans. Inf. Theory, vol. 52, no. 4, pp. 1289–1306, 2006.
  • [3] Y. C. Eldar and G. Kutyniok, ”Compressed Sensing: Theory and Applications”, Cambridge University Press, May 2012.
  • [4] W. Pu, X. Wang, J. Wu, Y. Huang and J. Yang, “A Novel Strategy for Video SAR Imaging Based on low-rank tensor recovery,” IEEE Trans. Neural Netw. Learn. Syst., vol. 32, no. 1, pp. 188–202, 2021.
  • [5] W. Pu and J. Wu, ”OSRanP: A Novel Way for Radar Imaging Utilizing Joint Sparsity and Low-Rankness,” IEEE Trans. Computational Imaging, vol. 6, pp. 868-882, 2020.
  • [6] N. Anantrasirichai, W. Hayes, M. Allinovi, D. Bull and A. Achim, ”Line Detection as an Inverse Problem: Application to Lung Ultrasound Imaging,” IEEE Trans. Medical Imag., vol. 36, no. 10, pp. 2045-2056, Oct. 2017
  • [7] M. Azghani, P. Kosmas and F. Marvasti, ”Microwave Medical Imaging Based on Sparsity and an Iterative Method With Adaptive Thresholding,” IEEE Trans. Medical Imag., vol. 34, no. 2, pp. 357-365, Feb. 2015
  • [8] P. Munger, G. R. Crelier, T. M. Peters and G. B. Pike, ”An inverse problem approach to the correction of distortion in EPI images,” IEEE Trans. Medical Imag., vol. 19, no. 7, pp. 681-689, July 2000
  • [9] E. J. Candes, J. Romberg, and T. Tao, “Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information,” IEEE Trans. Inf. Theory, vol. 52, no. 2, pp. 489–509, 2006.
  • [10] L. I. Rudin, S. Osher and E. Fatemi, ”Nonlinear total variation based noise removal algorithms,” Physica D, vol. 60, pp. 259-268, 1992.
  • [11] A. N., Tikhonov, V. Y. Arsenin, ”Solutions of Ill-Posed Problems”, 1977, New York: Winston.
  • [12] R. Tibshirani, ”Regression Shrinkage and Selection via the Lasso,” Journal of the Royal Statistical Society, Series B (Methodological), vol. 73, pp. 267-288, 1996.
  • [13] K. H. Jin, M. T. McCann, E. Froustey and M. Unser, ”Deep Convolutional Neural Network for Inverse Problems in Imaging,” IEEE Trans.Signal Process.,, vol. 26, no. 9, pp. 4509-4522, Sept. 2017
  • [14] G. Ongie, A. Jalal, C. A. Metzler, R. G. Baraniuk, A. G. Dimakis and R. Willett, ”Deep Learning Techniques for Inverse Problems in Imaging,” IEEE Journal on Selected Areas in Information Theory, vol. 1, no. 1, pp. 39-56, May 2020
  • [15] N. Chodosh and S. Lucey, ”When to Use Convolutional Neural Networks for Inverse Problems,” IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, pp. 8223-8232, 2020.
  • [16] Chao Dong, Chen Change Loy, Kaiming He and Xiaoou Tang, ”Image super-resolution using deep convolutional networks”, IEEE Trans. Pattern Anal. Mach. Intell., vol. 38, no. 2, pp. 295-307, 2016.
  • [17] K. Gregor and Y. LeCun, “Learning fast approximations of sparse coding,” in Proc. 27th Int. Conf. Mach. Learn. (ICML), pp. 399–406, 2010.
  • [18] V. Monga, Y. Li, Y. C. Eldar. “Algorithm Unrolling: Interpretable, Efficient Deep Learning for Signal and Image Processing.” IEEE Signal Process. Magazine, vol. 38, no. 2, pp. 18–44, 2021.
  • [19] Y. Yang, J. Sun, H. LI, and Z. Xu, “ADMM-CSNet: A Deep Learning Approach for Image Compressive Sensing,” IEEE Trans. Pattern Anal. Mach. Intell., pp. 1–1, to appear, 2019.
  • [20] Y. Li, M. Tofighi, J. Geng, V. Monga and Y. C. Eldar, ”Efficient and Interpretable Deep Blind Image Deblurring Via Algorithm Unrolling,” IEEE Trans. Comp. Imag., vol. 6, pp. 666-681, 2020.
  • [21] Y. Chen and T. Pock, “Trainable Nonlinear Reaction Diffusion: A Flexible Framework for Fast and Effective Image Restoration,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 39, no. 6, pp. 1256–1272, 2017.
  • [22] D. Hyun et al., ”Deep Learning for Ultrasound Image Formation: CUBDL Evaluation Framework &\& Open Datasets,” IEEE Transactions on Ultrasonics, Ferroelectrics, and Frequency Control, early access.
  • [23] G. D. Yoffe and Y. C. Eldar, ”Learned SPARCOM: unfolded deep super-resolution microscopy,” Opt. Express, vol. 28, pp. 27736-27763, 2020.
  • [24] M. A. Herman and T. Strohmer, ”General Deviants: An Analysis of Perturbations in Compressed Sensing,” IEEE J. Sel. Top. Signal Process., vol. 4, no. 2, pp. 342-349, April 2010
  • [25] Y. Chi, L. L. Scharf, A. Pezeshki and A. R. Calderbank, ”Sensitivity to Basis Mismatch in Compressed Sensing,” IEEE Trans. Signal Process., vol. 59, no. 5, pp. 2182-2195, May 2011
  • [26] H. Zhu, G. Leus, G. B. Giannakis, ”Sparsity-cognizant total least-squares for perturbed compressive sampling”, IEEE Trans.Signal Process., vol. 59, no. 5, pp. 2002–2016, 2011.
  • [27] J. T. Parker, V. Cevher, P. Schniter, ”Compressive sensing under matrix uncertainties: Anapproximate message passing approach”,in Proceedings of Asilomar Conference Signals, Systems and Computers, Pacific Grove, CA, USA, 2011, pp. 804–808.
  • [28] S. G. Lingala and M. Jacob, ”Blind Compressive Sensing Dynamic MRI,” IEEE Trans. Medi. Imag., vol. 32, no. 6 pp. 1132-1145, 2013.
  • [29] W. Pu, J. Wu, X. Wang, Y. Huang, Y. Zha, and J. Yang, Joint Sparsity-Based Imaging and Motion Error Estimation for BFSAR, IEEE Trans. Geosci. Remote Sens., vol. 57, no. 3, pp. 1393-1408, 2019.
  • [30] P. Song, X. Deng, J. F. C. Mota, N. Deligiannis, P. L. Dragotti and M. R. D. Rodrigues, ”Multimodal Image Super-Resolution via Joint Sparse Representations Induced by Coupled Dictionaries,” IEEE Trans. Comp. Imag., vol. 6, pp. 57-72, 2020.
  • [31] Szegedy, Christian, Zaremba, Wojciech, Sutskever, Ilya, Bruna, Joan, Erhan, Dumitru, Goodfellow, Ian, and Fergus, Rob, ”Intriguing properties of neural networks,” ICLR, 2014b.
  • [32] Martin Abadi, Andy Chu, Ian Goodfellow, H Brendan McMahan, Ilya Mironov, Kunal Talwar, and Li Zhang, Andy Chu, Ian Goodfellow, H Brendan McMahan, Ilya Mironov, Kunal Talwar, and Li Zhang, ”Deep learning with differential privacy,” in Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security. ACM, 308–318.
  • [33] A. N. Bhagoji, D. Cullina, C. Sitawarin and P. Mittal, ”Enhancing robustness of machine learning systems via data transformations,” in 2018 52nd Annual Conference on Information Sciences and Systems (CISS), 2018, pp. 1-5
  • [34] Maya Kabkab Pouya Samangouei and Rama Chellappa, ”Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models,” arXiv preprint, arXiv:1805.06605 (2018).
  • [35] S. Gu and L. Rigazio, ”Towards Deep Neural Network Architectures Robust to Adversarial Examples”, arXiv preprint, arXiv:1412.5068 (2014).
  • [36] W. Pu, C. Zhou, Y. C. Eldar and M. R. D. Rodrigues, ”REST: Robust lEarned Shrinkage-Thresholding Network Taming Inverse Problems with Model Mismatch,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2885-2889, 2021
  • [37] L. Zhu, X. Wu, Z. Sun, Z. Jin, and E. Weiland, et al., “Compressedsensing accelerated 3-dimensional magnetic resonance cholangiopancreatography: Application in suspected pancreatic diseases,” Invest. Radiol., vol. 53, no. 3, pp. 150–157, 2018.
  • [38] Y. Hitomi, J. Gu, M. Gupta, T. Mitsunaga, and S. K. Nayar, “Video from a single coded exposure photograph using a learned overcomplete dictionary,” in Proc. IEEE Int. Conf. Comput. Vis., 2011, pp. 287–294.
  • [39] M. F. Duarte, M. A. Davenport, D. Takhar, J. N. Laska, T. Sun, K. F. Kelly, and R. G. Baraniuk, “Single-pixel imaging via compressive sampling,” IEEE Signal Process. Mag., vol. 25, no. 2, pp. 83–91, Mar. 2008.
  • [40] W. Pu, J. Wu, ”OSRanP: A Novel Way for Radar Imaging Utilizing Joint Sparsity and Low-Rankness,” IEEE Trans. Comp. Imag., vol. 6, pp. 868-882, 2020.
  • [41] Tibshirani, and J. Ryan, ”The Lasso Problem and Uniqueness,” Electronic Journal of Statistics vol. 7, no. 1, pp. 1456-1490, 2013.
  • [42] B. Amir, and M. Teboulle, ”A fast iterative shrinkage-thresholding algorithm for linear inverse problems,” SIAM journal on imaging sciences, vol. , no. 1, pp. 183-202, 2009.
  • [43] S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eckstein, ”Distributed optimization and statistical learning via the alternating direction method of multipliers,” Found. Trends Mach. Learn., vol. 3, no. 1, pp. 1-122, 2011.
  • [44] W. Dinkelbach, “On nonlinear fractional programming,” Manag. Sci., vol. 13, no. 7, pp. 492-498, Mar. 1967.
  • [45] I. G. Akrotirianakis and C. A. Floudas, “Computational experience with a new class of convex underestimators: Box-constrained NLP problems,” J. Global Optim., vol. 29, no. 3, pp. 249-264, Jul. 2004.
  • [46] R. Arablouei, S. Werner, K. Dogancay, “Analysis of the gradient-decent total least-squares algorithm,” IEEE Trans. Signal Process., vol. 62, no. 5, pp. 1256-1264, 2014.
  • [47] Chen S S, Donoho D L, ”Saunders MA. Atomicdecomposition by basis pursuit,” SIAM review, vol. 43, no. 1, pp 129-159, 2001.
  • [48] A. Ben-Tal, L. E. Ghaoui, A. Nemirovski, ”Robust Optimization”, Princeton University Press, Aug 2009.