This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

MF-JMoDL-Net: A Deep Network for Azimuth Undersampling Pattern Design and Ambiguity Suppression for Sparse SAR Imaging

Yuwei Wua,b,c,d,e, Zhe Zhang b,c,d,e,*, Xiaolan Qiub,c,d,e, Yao Zhaof, Weidong Yua,b,e a Department of Space Microwave Remote Sensing System, Beijing, China.
b Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing, China.
c Suzhou Key Laboratory of Microwave Imaging, Processing and Application Technology, Suzhou, Jiangsu, China
d Suzhou Aerospace Information Research Institute, Suzhou, Jiangsu, China
e School of Electronic, Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing, China.
f Guangdong University of Technology, Guangzhou, Guangdong, China.
* Corresponding author: Zhe Zhang, Email: [email protected].
Abstract

Traditionally, the range swath of a synthetic aperture radar (SAR) system is constrained by its pulse repetition frequency (PRF). Given the system complexity and resource constraints, it is often difficult to achieve high imaging performance and low ambiguity without compromising the swath. In this paper, we propose a joint optimization framework for sparse strip SAR imaging algorithms and azimuth undersampling patterns based on a deep convolutional neural network, combined with matched filter (MF) approximate measurement operators and inverse MF operators, referred to as MF-JMoDL-Net, for sparse SAR imaging methods. Compared with conventional sparse SAR imaging, MF-JMoDL-Net enables us to alleviate the limitations imposed by PRF. In the proposed scheme, joint and continuous optimization of azimuth undersampling patterns and convolutional neural network parameters are implemented to suppress azimuth ambiguity and enhance sparse SAR imaging quality. Experiments and comparisons under various conditions demonstrate the effectiveness and superiority of the proposed framework in imaging results.

keywords:
Azimuth ambiguity suppression, deep learning, undersampling, sparse imaging, synthetic aperture radar (SAR)

1 Introduction

Synthetic Aperture Radar (SAR) is a prominent microwave imaging technology, renowned for its all-weather, round-the-clock, and high-resolution capabilities. It has been extensively utilized in remote sensing and monitoring applications [1]. Both contemporary civilian and military applications demand high spatial resolution. In accordance with the Nyquist sampling theorem, a high pulse repetition frequency (PRF) is necessary to attain a high azimuth resolution. Regrettably, an elevated PRF escalates the volume of echo data, potentially exceeding the on-board storage capacity and downlink bandwidth of a satellite. Consequently, an efficient azimuth undersampling scheme that minimizes azimuth ambiguity is of paramount importance for SAR systems.

To mitigate azimuth ambiguity, various signal processing methods have been proposed. The in-phase cancellation technique introduced in [2] alters the phase of the ambiguity signal and reduces azimuth ambiguity by constructing a correction filter. The restoration method [3] employs a defined metric, SAR system parameters, and the local mean energy of SAR images to identify ambiguity pixels, followed by selecting an ambiguity region restoration mechanism based on the size of ambiguity regions. The inpainting method proposed in [4] combines Wiener filtering and local azimuth ambiguity ratio (AASR) estimation to suppress N-order azimuth ambiguity. While this method can eliminate N-order azimuth ambiguity, it is unable to flexibly recover the actual targets affected by ambiguity. Nonetheless, all these methods focus on suppressing azimuth ambiguity in the context of uniform sampling rather than nonuniform sampling.

In contrast to uniform sampling, stochastic sampling has been introduced and investigated by several researchers [5, 6, 7]. Subsequently, stochastic sampling patterns were adopted in computer graphics for antialiasing, such as jittered sampling and Poisson disk sampling [8, 9]. Villano et al. were the first to introduce staggered sampling [10] into SAR imaging, termed staggered-SAR, where the PRF was continuously varied [11, 12, 13], and ambiguity in the image manifested as noise-like disturbances rather than localized artifacts [14]. Regrettably, this processing method necessitates the system to operate at a high azimuth oversampling rate, resulting in significantly increased data rates for SAR hardware systems.

Compressive sensing (CS) theory [15, 16] has been introduced into radar imaging [17, 18], enabling SAR imaging at low sampling rates (CS-SAR). Zhang et al. [19], He et al. [20], and Cetin et al. [21] proposed sparse SAR imaging, which reconstructs the scene of interest by solving a regularization problem using CS recovery algorithms, such as Orthogonal Matching Pursuit (OMP) [22], Compressive Sampling Matching Pursuit (CoSaMP) [23, 24], Fast Iterative Shrinkage-Thresholding Algorithm (FISTA) [25], and Iterative Shrinkage-Thresholding Algorithm (ISTA) [26]. These algorithms enable direct processing of downsampling raw data. However, in the context of satellite SAR imaging, handling a large measurement matrix while processing raw data is challenging. To address this limitation, an efficient regularization algorithm for SAR sparse imaging is proposed, which firstly suggested that the matched filtering (MF) algorithm could be considered an imaging operator, thus forming a new CS-SAR imaging method for high-quality, high-resolution imaging under sub-Nyquist rate sampling [27, 28]. With the support of sparse SAR imaging technology, the possibility of azimuth undersampling has garnered researchers’ attention. Sun et al. [29] proposed a CS-SAR imaging approach based on azimuth Poisson disk-like nonuniform sampling, which distinguishes itself from staggered-SAR by adopting a sampling pattern based on Poisson downsampling rate instead of linear oversampling rate. This method significantly reduces the computational complexity of CS reconstruction from 𝒪(2MN)\mathcal{O}(2MN) to 𝒪(4NlogN)\mathcal{O}(4N\log N), but it only corrects the range walk during the range cell migration correction (RCMC) process without addressing the range curvature. To resolve this issue, [30, 31] proposed an algorithm that combines the CSA operator and ISTA to correct both the range walk and the range curvature under Poisson disk nonuniform downsampling conditions. Although sparse SAR imaging offers the aforementioned advantages, it also introduces nonuniform sampling, echo loss, and non-ideal AAP issues, leading to considerable aliasing and ambiguity particularly in azimuth direction.

Deep learning has achieved remarkable performance in signal and image processing, with the deep unrolling algorithm demonstrating the potential for network-based SAR imaging reconstruction [32]. Mason et al. [33] introduced a recurrent autoencoder network architecture based on the ISTA that incorporated SAR modeling, exhibiting faster convergence and reduced reconstruction error compared to standard iterative analytic methods. Pu et al. [34] proposed an autoencoder-structured deep SAR imaging algorithm, where motion compensation is considered to mitigate the impact of motion errors on imaging results. In [35], an end-to-end SAR deep learning imaging method based on a sparse optimization iterative algorithm was suggested to enhance the universality and generalization ability of the imaging method for SAR echo data. Li et al. [36] developed a target-oriented SAR imaging model, where generalized regularization was employed to characterize target features, contributing to an improved signal-to-clutter ratio (SCR) in the reconstructed image. It is evident that existing deep network-based SAR imaging methods primarily concentrate on enhancing the quality of SAR images. However, no research combines the deep network and azimuth sampling pattern in sparse SAR imaging. Current deep network-based sparse SAR imaging works typically directly adopt canonical uniform or random undersampling schemes, but these undersampling patterns are suboptimal in most cases. In the biomedical imaging field, Aggarwal et al. [37] proposed a joint model-based deep learning approach for optimizing downsampling and reconstruction (J-MoDL) specifically for magnetic resonance imaging (MRI), based on an image reconstruction framework with a CNN-based regularization prior [38]. This method significantly enhances imaging quality and the performance of most deep learning reconstruction algorithms.

Inspired by J-MoDL, this paper proposes a joint optimization framework for sparse strip SAR imaging algorithms and azimuth undersampling patterns based on a deep convolutional neural network combined with an MF approximate measurement operator. The main contributions of this article are as follows:

1) A learnable undersampling pattern framework is proposed: This is the first time a deep learning model has been introduced into the optimization of SAR azimuth sampling patterns. In this paper, we design a data-driven azimuth sampling model that extracts features using the U-Net network [39]. Leveraging the exceptional information extraction capability of deep neural networks, effective information is separated from clutter information, which contributes to obtaining more information of interest while using as few pulses as possible. The proposed sampling method outperforms existing azimuth pulse repetition interval (PRI) designs in terms of imaging performance.

2) A deep network imaging approach based on nonuniform MF operator is presented: The MF-JMoDL-Net utilizes an approximate measurement operator to significantly reduce storage and computation costs. We have originally incorporated the nonuniform fast Fourier transform (NUFFT) MF operator into the network-based SAR imaging framework, making the proposed method efficient and feasible. Furthermore, the backpropagation based on conjugate gradients can enforce data consistency, unlike conventional steepest descent updates. Its ability to work with smaller CNN modules allows it to learn from smaller datasets, making it well-suited to the issue of small SAR imaging training datasets in deep learning applications.

3) Experiments are conducted under various conditions: To verify the effectiveness of the proposed method, we make comparisons among the proposed algorithm, Poisson-disk, and staggered sampling patterns, and perform experiments on conventional azimuth ambiguity suppression. All the results demonstrate the feasibility and reliability of the proposed method. Additionally, we investigate the preliminary analysis of the obtained sampling patterns for different scenes, and explore the imaging performance at 50%, 25%, and 12.5% undersampling rates, aiming to discuss how this novel SAR imaging method performs on ambiguity suppression as the sampling rate decreases.

The rest of this paper is organized as follows: Section 2 presents the SAR signal model and the proposed imaging model. Section 3 describes the solution of the imaging model and provides details on the MF-JMoDL-Net. In Section 4, experimental results and related analysis are presented. Finally, conclusions are drawn in Section 5.

Refer to caption
Figure 1: Geometric configuration of side view stripmap mode SAR.

2 Problem Setting and Approaches

In this section, we first provide the stripmap SAR signal model. Then, in Subsection 2.2, the joint optimized sampling scheme and reconstruction framework are presented to clarify the intention of the proposed SAR imaging model. Finally, in Subsection 2.3, the nonuniform MF-based approximate measurement model is introduced for the completeness of this article.

2.1 Stripmap SAR Signal Model

We assume that the SAR system operates in single-channel strip mode. The geometric configuration of a typical side-looking SAR is shown in Fig.1 Suppose that the platform carrying the imaging radar moves in a straight line relative to the observation scene at a constant speed vv. For stripmap SAR, the direction of the antenna beam remains fixed throughout the entire movement. Consequently, the distance between the phase center of the antenna and the target point with coordinates (x,y)(x,y) in the observation scene is

R(η,x,y)=(xvη)2+R02R\left({\eta,x,y}\right)=\sqrt{(x-v\eta)^{2}+R_{0}^{2}} (1)

where τ\tau and η\eta are the fast time and the slow time, respectively, xx and yy are the spatial coordinates of the target point along the azimuth and range directions, and R0R_{0} is the closest range between target and platform.

The signal transmitted by the antenna to the observation scene is a linear frequency modulated signal (LFM), therefore, the echo received by the imaging radar of the target point (x,y)(x,y) in the observation scene is

s(τ,η,x,y)=A0σ(x,y)ωr(τ2R(η,x,y)c)exp{jπKr(τ2R(η,x,y)c)2}ωa(ηηc)exp{j4πf0R(η,x,y)c}\begin{split}s\left({\tau,\eta,x,y}\right)=A_{0}\sigma\left({x,y}\right)\cdot\omega_{r}\left({\tau-\frac{2R\left({\eta,x,y}\right)}{c}}\right)\\ \cdot exp\left\{{j\pi K_{r}\left({\tau-\frac{2R\left({\eta,x,y}\right)}{c}}\right)^{2}}\right\}\cdot\omega_{a}\left({\eta-\eta_{c}}\right)\\ \cdot{\mathit{\exp}\left\{{-\frac{j4\pi f_{0}R\left({\eta,x,y}\right)}{c}}\right\}}\end{split} (2)

where A0A_{0} is the observation gain determined by the antenna; σ(x,y)\sigma\left({x,y}\right) is the backscattering coefficient of the target point (x,y)(x,y); ωa()\omega_{a}(\cdot) and ωr()\omega_{r}(\cdot) are the azimuth and range signal envelope, respectively; f0f_{0} is the carrier frequency; cc is the speed of light. The SAR echo reflected from the scatterers generated by the antenna beam coverage area Ω\Omega can be expressed as

s(τ,η)=x,yΩs(τ,η,x,y)𝑑x𝑑y+N(τ,η)\begin{split}s\left({\tau,\eta}\right)={\iint_{x,y\in\Omega}{s\left({\tau,\eta,x,y}\right)dxdy}}+N\left({\tau,\eta}\right)\end{split} (3)

where N(τ,η)N(\tau,\eta) is the additive measurement noise. By discretizing the observation scene Ω\Omega into Mr×MaM_{r}\times M_{a} grids and sampling the echo signal into a matrix, the discrete echo can be expressed as

s(τnr,ηna)=mrMrmaMah[τnr,ηna,xmr,yma]σ(xmr,yma)+N(τnr,ηna)\begin{split}s\left({\tau_{n_{r}},\eta_{n_{a}}}\right)={\sum\limits_{m_{r}}^{M_{r}}{\sum\limits_{m_{a}}^{M_{a}}h\left[\tau_{n_{r}},\eta_{n_{a}},x_{m_{r}},y_{m_{a}}\right]}}\\ \cdot{\sigma\left({x_{m_{r}},y_{m_{a}}}\right)}+N\left(\tau_{n_{r}},\eta_{n_{a}}\right)\end{split} (4)

where s(τnr,ηna)s\left({\tau_{n_{r}},\eta_{n_{a}}}\right) is the nrn_{r}-th range sample at the nan_{a}-th pulse of raw signal s(τ,η)s\left({\tau,\eta}\right) , MrM_{r} and MaM_{a} are the numbers of the discrete cells in the range and azimuth directions, respectively. The h[τnr,ηna,xmr,yma]h\left[\tau_{n_{r}},\eta_{n_{a}},x_{m_{r}},y_{m_{a}}\right] is defined by

h[τnr,ηna,xmr,yma]=ωr(τnr2R(ηna,xmr,yma)c)exp{jπKr(τnr2R(ηna,xmr,yma)c)2}ωa(ηnaηc)exp{j4πf0R(ηna,xmr,yma)c}\begin{split}h\left[{\tau_{n_{r}},\eta_{n_{a}},x_{m_{r}},y_{m_{a}}}\right]=\omega_{r}\left({\tau_{n_{r}}-\frac{2R\left({\eta_{n_{a}},x_{m_{r}},y_{m_{a}}}\right)}{c}}\right)\\ \cdot exp\left\{{j\pi K_{r}\left({\tau_{n_{r}}-\frac{2R\left({\eta_{n_{a}},x_{m_{r}},y_{m_{a}}}\right)}{c}}\right)^{2}}\right\}\\ \cdot\omega_{a}\left({\eta_{n_{a}}-\eta_{c}}\right)\cdot{\mathit{\exp}\left\{{-\frac{j4\pi f_{0}R\left({\eta_{n_{a}},x_{m_{r}},y_{m_{a}}}\right)}{c}}\right\}}\end{split} (5)

Write the discretized data s(τnr,ηna)s\left({\tau_{n_{r}},\eta_{n_{a}}}\right) and the reflectivity map σ(xmr,yma)\mathbf{\sigma}\left({x_{m_{r}},y_{m_{a}}}\right) as two matrices: 𝐒M×Nr\mathbf{S}\in\mathbb{C}^{{M\times N}_{r}} and σMa×Mr\mathbf{\sigma}\in\mathbb{C}^{{M_{a}\times M}_{r}}, respectively, where MM is the number of transmitted pulses in the azimuth direction; NrN_{r} is the number of samples in the range direction. Then, (4) can be rewritten as

𝐒=𝐇σ+𝐧\begin{split}\mathbf{S}=\mathbf{H}\mathbf{\sigma}+\mathbf{n}\end{split} (6)

where 𝐧\mathbf{n} is the vector of additive noise, and 𝐇\mathbf{H} is defined by

𝐇=[a[τ1,η1,x1,y1]a[τ1,η1,xMa,yMr]a[τ1,η2,x1,y1]a[τ1,η2,xMa,yMr]a[τ1,ηM,x1,y1]a[τ1,ηM,xMa,yMr]a[τNr,ηM,x1,y1]a[τNr,ηM,xMa,yMr]].\mathbf{H}=\begin{bmatrix}{a\left[\tau_{1},\eta_{1},x_{1},y_{1}\right]}&\cdots&{a\left[\tau_{1},\eta_{1},x_{M_{a}},y_{M_{r}}\right]}\\ {a\left[\tau_{1},\eta_{2},x_{1},y_{1}\right]}&\cdots&{a\left[\tau_{1},\eta_{2},x_{M_{a}},y_{M_{r}}\right]}\\ \vdots&\ddots&\vdots\\ {a\left[\tau_{1},\eta_{M},x_{1},y_{1}\right]}&\cdots&{a\left[\tau_{1},\eta_{M},x_{M_{a}},y_{M_{r}}\right]}\\ \vdots&\ddots&\vdots\\ {a\left[\tau_{N_{r}},\eta_{M},x_{1},y_{1}\right]}&\cdots&{a\left[\tau_{N_{r}},\eta_{M},x_{M_{a}},y_{M_{r}}\right]}\\ \end{bmatrix}. (7)

Conventionally, the intention of CS-SAR imaging is to reconstruct σ\mathbf{\sigma} from the measurements 𝐒\mathbf{S}.

2.2 MF-Based Image Formation and Proposed Forward Model

In this subsection, we will introduce and discuss the typical MF-Based CS-SAR imaging model to explain the original intention of the proposed method. Then, we discuss the forward model with the sampling pattern, which is the central assumption of the proposed MF-JMoDL-Based SAR imaging model.

1) Typical MF-Based CS-SAR Imaging Model: in the typical CS-SAR imaging model, the measurement vector 𝐇\mathbf{H} is generally resampled by an appropriate sampling matrix Θ\Theta. When the observation scene is sparse and the observation matrix 𝐀=Θ𝐇\mathbf{A}=\Theta\mathbf{H} satisfies the RIP condition [15], σ\mathbf{\sigma} can be exactly recovered from 𝐒\mathbf{S} with the Lq(0<q<1)L_{q}(0<q<1) optimization:

minσσqs.t.𝐒=𝐀σ\begin{split}{\min\limits_{\mathbf{\sigma}}\left\|\mathbf{\sigma}\right\|_{q}}~{}~{}~{}~{}~{}~{}~{}s.t.~{}~{}\mathbf{S}=\mathbf{A}\mathbf{\sigma}\end{split} (8)

To solve this problem, an equivalent regularization scheme with the following optimization problem can be used:

min𝜎{𝐒𝐀σ22+λσqq}\begin{split}\underset{\mathbf{\sigma}}{min~{}}\left\{{\left\|{\mathbf{S}-\mathbf{A}\mathbf{\sigma}}\right\|_{2}^{2}+\lambda\left\|\mathbf{\sigma}\right\|_{q}^{q}}\right\}\end{split} (9)

where λ\lambda is a regularization parameter. Generally speaking, the solution methods for (8) and (9) can be divided into two categories: greedy methods and iterative methods. However, regardless of the chosen method, the computational and storage costs are enormous due to the size of the measurement matrix 𝐇\mathbf{H}, which significantly reduces computational efficiency. Hence, the corresponding reconstruction algorithms are considerably more time-consuming than traditional matched filter MF-based focusing methods.

To address this issue, the MF-based sparse SAR imaging model [28, 27] offers a general principle on how the observation can be remodeled and approximated by any high-precision imaging procedure:

𝐆=𝐌1𝐇\begin{split}\mathbf{G}=\mathbf{M}^{-1}\approx\mathbf{H}\end{split} (10)

where 𝐌\mathbf{M} is the traditional MF imaging procedure that can be calculated through decoupling it into a series of 1-D operators in the frequency domain, like chirp scaling algorithm (CSA) and range doppler algorithm (RDA), 𝐆\mathbf{G} is any generalized right inverse of 𝐌\mathbf{M}.

2) Forward Model with Sampling Pattern: To optimize the azimuth sampling scheme of MF-based SAR imaging, we employ the model-based imaging algorithms [40][41]. The model-based imaging schemes use a continuous function of 𝐀\mathbf{A}, denoted by the operator HΘa()H_{\Theta_{a}}^{*}(\cdot), which maps σ\mathbf{\sigma} to 𝐒\mathbf{S}, and Θa\Theta_{a} represents the azimuth sampling pattern. Then, we construct the forward model as follows:

𝐒=HΘa(σ)\begin{split}\mathbf{S}=H_{\Theta_{a}}^{*}\left(\mathbf{\sigma}\right)\end{split} (11)

Due to the recovery of σ\mathbf{\sigma} from 𝐒\mathbf{S} is ill-posed, model-based imaging algorithms pose the recovery as a regularized optimization scheme:

σ=argminσ𝐒HΘa(σ)22dataconsistency+λN(σ)regularization\begin{split}\mathbf{\sigma}={\arg{\min\limits_{\mathbf{\sigma}}{\underset{\rm data~{}consistency}{\underbrace{\left\|{\mathbf{S}-H_{\Theta_{a}}^{*}\left(\mathbf{\sigma}\right)}\right\|_{2}^{2}}}+\lambda\underset{\rm regularization}{\underbrace{N\left(\mathbf{\sigma}\right)}}}}}\end{split} (12)

here N(σ)N\left(\mathbf{\sigma}\right) is a regularization penalty. When σ\mathbf{\sigma} is an artifact-free image, N(σ)N\left(\mathbf{\sigma}\right) is a small scalar, otherwise, it is a high scalar for noisy images. Commonly-used regularizers in imaging reconstruction include total variation regularization [42], transform domain sparsity [43], and structured low-rank methods [44]. These regularizers aim to enforce specific desirable properties on the reconstructed image, such as smoothness, sparsity in certain domains, or low-rank structures, ultimately leading to improved image quality and reduced artifacts.

2.3 Nonuniform MF Operator

Refer to caption
(a) CSA operator
Refer to caption
(b) Inverse CSA operator
Figure 2: Procedure of CSA operator and inverse CSA operator.

In order to address the challenges posed by the proposed model in 2.2, two main problems must be overcome.

Firstly, the large-scale exact measurement matrix 𝐀\mathbf{A} presents significant computational and storage requirements, which would reduce the efficiency of the proposed SAR imaging model. To tackle this issue, the MF-based SAR imaging method is employed to construct an approximate measurement operator by decoupling azimuth and range processing. Following a similar approach to the work in [28], the chirp scaling algorithm (CSA) is used to build the approximate measurement operator.

Secondly, the sampling pattern proposed in this paper leads to nonuniform samples in the azimuth direction. Consequently, a nonuniform MF operator HΘa()H_{\Theta_{a}}(\cdot) is needed, which is different from the regular MF operator. In a similar manner to the work in [31], the chirp scaling algorithm (CSA) is adopted to construct the approximate measurement operator. The process for constructing the approximate measurement operator is illustrated in Fig.2(a) and Fig.2(b).

By addressing these two challenges, the proposed SAR imaging model can be solved efficiently while maintaining the benefits of the optimized azimuth sampling pattern.

The formulation of CSA can be expressed as follows:

σ^=𝐈𝐅η(𝐈𝐅τ(𝐅τ(𝐍𝐅ηΘa(𝐒)θ1)θ2)θ3)\begin{split}\hat{\mathbf{\sigma}}=\mathbf{I}{\mathbf{F}_{\eta}}\left({\mathbf{I}{\mathbf{F}_{\tau}}\left({{\mathbf{F}_{\tau}}\left({\mathbf{N}{\mathbf{F}_{{{\eta}_{{{\rm{\Theta}}_{a}}}}}}\left(\mathbf{S}\right)\odot{{\theta}_{1}}}\right)\odot{{\theta}_{2}}}\right)\odot{{\theta}_{3}}}\right)\end{split} (13)

In the given context, σ^\hat{\mathbf{\sigma}} represents the reconstructed SAR image matrix, while 𝐅τ\mathbf{F}{\tau} and 𝐅η\mathbf{F}{\eta} denote the fast Fourier transform (FFT) along the range direction and azimuth direction, respectively. 𝐈𝐅τ\mathbf{IF}{\tau} is the inverse fast Fourier transform (IFFT) along the range direction. The phase terms for RCMC, range compression, azimuth-range decoupling, and azimuth compression are represented by filters θ1\theta{1}, θ2\theta_{2}, and θ3\theta_{3}, respectively. The expressions for these phase terms can be found in [31], and as such, they are not repeated here.

The equivalent CSA operator is denoted as HΘa()H_{\Theta_{a}}(\cdot). Therefore, equation (13) can be expressed as σ^=HΘa(𝐒)\hat{\mathbf{\sigma}}=H_{\Theta_{a}}\left(\mathbf{S}\right). This formulation allows for efficient processing of the SAR imaging problem while taking into account the optimized azimuth sampling pattern.

Now we defined the nonuniform Fourier transform matrix 𝐍𝐅ηΘa\mathbf{NF}_{\eta_{\Theta_{a}}} in (10), which is used to transform the raw data into the range–doppler domain as

𝐍𝐅ηΘa=[α1,α2,,α𝐌]\begin{split}{\mathbf{N}\mathbf{F}}_{\eta_{\Theta_{a}}}=\left[{\mathbf{\alpha}_{1},\mathbf{\alpha}_{2},\ldots,\mathbf{\alpha}_{\mathbf{M}}}\right]\end{split} (14)
Refer to caption
Figure 3: The recursive framework that alternates between CNN based denoiser block DΦD_{\Phi} in (25) and the data-consisitency (DC) layer in (26).

it consists of MM column vectors defined by

αm=[exp(j2πηmfη(1)),,exp(j2πηmfη(Na))]T\begin{split}{\alpha_{m}}={\left[{exp\left({-j2\pi{\eta_{m}}f_{\eta}^{\left(1\right)}}\right),\ldots,exp\left({-j2\pi{\eta_{m}}f_{\eta}^{\left({{N_{a}}}\right)}}\right)}\right]^{T}}\end{split} (15)

In equation (15), ηm\eta_{m} represents the mm-th sample in the azimuth direction, with m=1,2,,Mm=1,2,\dots,M, where MM is the number of samples in azimuth. The nn-th sample in Doppler frequency is given by fη(n)=(n1)PRF/Naf_{\eta}^{(n)}=(n-1)PRF/N_{a}, where n=1,2,,Nan=1,2,\dots,N_{a}, and NaN_{a} is the number of uniform frequency grids in the Doppler domain. PRFPRF denotes the azimuth pulse repetition frequency.

By applying the CSA operator to the raw data 𝐒\mathbf{S}, we can obtain the reconstructed backscattering coefficients of the scatterers in the observed scene. These coefficients represent the true backscattering coefficients we desire, which are approximately equal to σ\mathbf{\sigma}. The expression is as follows:

HΘa(𝐒)=σ^σ\begin{split}H_{\Theta_{a}}\left(\mathbf{S}\right)=\hat{\mathbf{\sigma}}\approx\mathbf{\sigma}\end{split} (16)

From (13), it is evident that the CSA operator is invertible, as mentioned in [31]. By applying the inverse of the CSA operator to the true backscattering coefficients of the scene, we can obtain an approximation of the raw data 𝐒\mathbf{S}. The inverse CSA operator can be explicitly expressed as follows:

𝐒^=𝐍𝐈𝐅ηΘa(𝐈𝐅τ(𝐅τ(𝐅η(σ)θ3)θ2)θ1)\begin{split}\hat{\mathbf{S}}={\mathbf{N}\mathbf{I}\mathbf{F}}_{\eta_{\Theta_{a}}}\left(\mathbf{I}\mathbf{F}_{\tau}\left(\mathbf{F}_{\tau}\left(\mathbf{F}_{\eta}\left(\mathbf{\sigma}\right)\odot\theta_{3}^{\mathbf{*}}\right)\odot\theta_{2}^{\mathbf{*}}\right)\odot\theta_{1}^{\mathbf{*}}\right)\end{split} (17)

where 𝐅η\mathbf{F}_{\eta} is the azimuth FFT operator. 𝐒^\hat{\mathbf{S}} is the approximated echo data matrix generated by HΘa()H_{\Theta_{a}}(\cdot), then the (17) can be expressed as 𝐒^=HΘa(σ)\hat{\mathbf{S}}=H_{\Theta_{a}}\left(\mathbf{\sigma}\right). the nonuniform inverse Fourier transform 𝐍𝐈𝐅ηΘa{\mathbf{N}\mathbf{I}\mathbf{F}}_{\eta_{\Theta_{a}}} composed of 𝐍𝐚\mathbf{N}_{\mathbf{a}} column vectors defined by

𝐍𝐈𝐅ηΘa=[β1,β2,,β𝐍𝐚]\begin{split}{\mathbf{N}\mathbf{I}\mathbf{F}}_{\eta_{\Theta_{a}}}=\left[{\mathbf{\beta}_{1},\mathbf{\beta}_{2},\ldots,\mathbf{\beta}_{\mathbf{N}_{\mathbf{a}}}\mathbf{~{}}}\right]\end{split} (18)

which is consist of 𝐍𝐚\mathbf{N}_{\mathbf{a}} column vectors defined by

β𝐧=1Na[exp(j2πη1fη(n)),,exp(j2πηMfη(n))]T\begin{split}\mathbf{\beta}_{\mathbf{n}}=\frac{1}{N_{a}}\left[exp\left({j2\pi\eta_{1}f_{\eta}^{(n)}}\right),\ldots,exp\left({j2\pi\eta_{M}f_{\eta}^{(n)}}\right)\right]^{T}\end{split} (19)

where η1,,ηM\eta_{1},\ldots,\eta_{M} are obtained according to proposed method.

3 Proposed Method

In this section, we aim to solve the proposed model-based SAR imaging framework from Subection 2.2. The process consists of four main parts:

1) Deriving a deep learning framework (Subsection 3.1): The deep learning framework is derived to facilitate the solution of the model-based SAR imaging framework. This will enable the efficient processing and optimization of SAR imaging using advanced deep learning techniques.

2) Joint optimization of sampling pattern (Subsection 3.2): We present the joint optimization of the azimuth sampling scheme in SAR imaging. This optimization allows the system to learn an effective sampling pattern for improved SAR imaging performance.

3) Unfolding the iterative solution into a deep network-based solution, MF-JMoDL-Net (Subsection 3.3): The iterative solution is unfolded into a deep network-based solution called MF-JMoDL-Net. This network will be capable of handling the SAR imaging problem with improved efficiency and performance compared to traditional iterative methods.

4) Loss function, backpropagation, and gradient calculation analysis (Subsection 3.4): In this part, we provide details on the loss function used for optimization, the backpropagation process for updating the network weights, and the gradient calculation. This information is crucial for the successful training and implementation of the MF-JMoDL-Net.

By addressing these four components, we will develop a comprehensive solution for the proposed model-based SAR imaging framework, leveraging the power of deep learning to improve the overall performance of SAR imaging systems.

3.1 JModel-Based Deep Learning Image Reconstruction

Consider the model proposed in 2.2, we define the regularizer as N(σ)=λ𝐓σ1N\left(\mathbf{\sigma}\right)=\lambda\left\|{\mathbf{T}\mathbf{\sigma}}\right\|_{\mathcal{L}_{1}} in transform domain sparsity, with Φ={𝐓,λ}\Phi=\left\{\mathbf{T},\lambda\right\} denoting the parameters of the transform and regularization. To denote the dependence on the regularization parameters as well as the sampling pattern of (12), we give the optimization problem of the form

σ^{Φ,Θa}=argmin𝜎𝐒HΘa(σ)22+NΦ(σ)\begin{split}\hat{\mathbf{\sigma}}\left\{{\Phi,\Theta}_{a}\right\}={\arg{\underset{\mathbf{\sigma}}{min~{}}{\left\|{\mathbf{S}-H_{\Theta_{a}}^{*}\left(\mathbf{\sigma}\right)}\right\|_{2}^{2}+}}}N_{\Phi}\left(\mathbf{\sigma}\right)\end{split} (20)

To address (13), we treat the image reconstruction as a regularization optimization problem and employ a deep learning approach rather than traditional fixed prior regularization to learn parameters from data samples with optimized azimuth sampling patterns, which are commonly referred to as data-driven methods. Classical data-driven image reconstruction techniques include direct inversion schemes [45][46] and the model-based deep learning framework (MoDL) [38]. Owing to the substantial reduction in the number of network parameters in MoDL, we focus on this deep learning framework, formulating its image recovery as follows:

σ^{Θa,Φ}=argmin𝜎𝐒HΘa(σ)22+σDΦ(σ)22\begin{split}\hat{\mathbf{\sigma}}\left\{{\Theta_{a},\Phi}\right\}={\arg{\underset{\sigma}{min~{}}{\left\|{\mathbf{S}-H_{\Theta_{a}}^{*}\left(\mathbf{\sigma}\right)}\right\|_{2}^{2}+\left\|{\mathbf{\sigma}-D_{\Phi}\left(\mathbf{\sigma}\right)}\right\|_{2}^{2}}}}\end{split} (21)

where DΦ(σ)D_{\Phi}\left(\mathbf{\sigma}\right) is the “denoised” version of σ\mathbf{\sigma}. To extract the ambiguity and noise in σ\mathbf{\sigma}, we let DΦD_{\Phi} be a residual learning-based CNN. Similar to [38], setting σn+Δσ=σ\mathbf{\sigma}_{n}+\mathrm{\Delta}\mathbf{\sigma}=\mathbf{\sigma}, the non-linear mapping DΦ(σn+Δσ)D_{\Phi}\left({\mathbf{\sigma}_{n}+\mathrm{\Delta}\mathbf{\sigma}}\right) can be approximated using Taylor series around the nthn^{th} iterate as

DΦ(σn+Δσ)DΦ(σn)zn+JnTx\begin{split}D_{\Phi}\left({\mathbf{\sigma}_{n}+\mathrm{\Delta}\mathbf{\sigma}}\right)\approx\underset{z_{n}}{\underbrace{D_{\Phi}\left(\mathbf{\sigma}_{n}\right)}}+J_{n}^{T}\nabla x\end{split} (22)

where JnJ_{n} is a Jacobian matrix, the penalty term can be approximated as

σDΦ(σn+Δσ)2σzn2+JnΔσ2\begin{split}\left\|{\mathbf{\sigma}-D_{\Phi}\left({\mathbf{\sigma}_{n}+\mathrm{\Delta}\mathbf{\sigma}}\right)}\right\|^{2}\approx\left\|{\mathbf{\sigma}-z_{n}}\right\|^{2}+\left\|{J_{n}\mathrm{\Delta}\mathbf{\sigma}}\right\|^{2}\end{split} (23)
Refer to caption
Figure 4: The unrolled architecture for KK iterations, which share weights with the denoising blocks DΦD_{\Phi}.

The above approximation is only valid in the vicinity of σn\mathbf{\sigma}_{n}, hence we obtain the alternating algorithm that approximate (21):

σn+1=(HΘaHΘa+I)1(zn+𝐒HΘaσ)\begin{split}\mathbf{\sigma}_{n+1}=\left({H_{\Theta_{a}}H_{\Theta_{a}}^{*}+I}\right)^{-1}\left({z_{n}+\mathbf{S}-H_{\Theta_{a}}^{*}\mathbf{\sigma}}\right)\end{split} (24)
zn+1=DΦ(σn+1)\begin{split}z_{n+1}=D_{\Phi}\left(\mathbf{\sigma}_{n+1}\right)\end{split} (25)

here, (24) is implemented using a conjugate gradient algorithm:

σn+1=(HΘaHΘa+I)1CΘa(HΘa(𝐒)+λzn)\begin{split}\mathbf{\sigma}_{n+1}=\underset{C_{\Theta_{a}}}{\underbrace{\left({H_{\Theta_{a}}H_{\Theta_{a}}^{*}+I}\right)^{-1}}}\left({H_{\Theta_{a}}\left(\mathbf{S}\right){+\lambda z}_{n}}\right)\end{split} (26)

This iterative algorithm is unrolled to obtain a deep recursive network MΘa,ΦM_{\Theta_{a},\Phi}, where the weights of the CNN blocks and data consistency blocks are shared across iterations, as shown in Fig. 3. Specifically, the solution to (21) is given by

σ^{Θa,Φ}=MΘa,Φ(HΘa(𝐒))\begin{split}\hat{\mathbf{\sigma}}\left\{{\Theta_{a},\Phi}\right\}=M_{\Theta_{a},\Phi}\left(H_{\Theta_{a}}\left(\mathbf{S}\right)\right)\end{split} (27)

Note that once unrolled, the image reconstruction algorithm is essentially a deep network, shown in Fig. 4. Thus, the main distinction between MoDL and direct-inversion scheme is the structure of network MΘa,ΦM_{\Theta_{a},\Phi}.

3.2 Joint Optimization of Sampling Pattern

In this work, we concentrate on optimizing the azimuth sampling pattern denoted by Θa\Theta_{a} in (11) and the parameter Φ\Phi of the reconstruction algorithm (20) to enhance the quality of the reconstructed images. Drawing inspiration from earlier sampling pattern learning approaches employed in MRI based on compressed sensing algorithms [47] [48], we incorporate these models into the SAR imaging field to optimize the azimuth sampling pattern Θa\Theta_{a}, such that

{Θa}=argminΘai=1Nσ^i{Θa,Φ}σi22\begin{split}\left\{\Theta_{a}^{*}\right\}={\arg{\min\limits_{\Theta_{a}}{\sum\limits_{i=1}^{N}\left\|{{\hat{\sigma}}_{i}\left\{{\Theta_{a},\Phi}\right\}-\sigma_{i}}\right\|_{2}^{2}}}}\end{split} (28)

is minimized. Here σ^i{Θa,Φ}{\hat{\sigma}}_{i}\left\{{\Theta_{a},\Phi}\right\} is the corresponding reconstruction of training image σi;i=1,,N\sigma_{i};~{}i=1,\ldots,N in the optimization.

To improve the reconstruction performance, this paper uses a MF operator model-based deep learning network (MF-JMoDL-Net) framework to jointly optimize both CΘaC_{\Theta_{a}} and DΦD_{\Phi} blocks in the MF-JMoDL framework (21), which refers to [37]. Given training data, this framework jointly learns the sampling pattern Θa\Theta_{a} and the CNN parameters Φ\Phi using

{Θa,Φ}=argminΘa,Φi=1NMΘa,Φ(HΘa(si))σi22\begin{split}\left\{{\Theta_{a}^{*},\Phi^{*}}\right\}={\arg{\min\limits_{\Theta_{a},\Phi}{\sum\limits_{i=1}^{N}\left\|{M_{\Theta_{a},\Phi}\left(H_{\Theta_{a}}^{*}\left(s_{i}\right)\right)-\sigma_{i}}\right\|_{2}^{2}}}}\end{split} (29)

where MΘa,ΦM_{\Theta_{a},\Phi} denotes a general deep learning network architecture that includes forward model as well as unrolled architectures denoted by (27).

3.3 MF-JMoDL-Net Architecture

Fig.3 illustrates the framework of the proposed MF-JMoDL, which alternates between CNN blocks DΦD_{\Phi} and data consistency blocks CΘaC_{\Theta_{a}} solely dependent on the sampling pattern. We unroll the MF-JMoDL framework for five iterations of alternating minimization to solve (21), as depicted in Fig.4. The implementation process of the forward operator HΘaH_{\Theta_{a}} is shown in Fig. 2(b), with the trainable parameters Θa=[ky1kyikyNa]T\Theta_{a}=\left[{k_{y_{1}}\ldots k_{y_{i}}\ldots k_{y_{N_{a}}}}\right]^{T}. Since the operator (HΘaHΘa+I)\left({H_{\Theta_{a}}H_{\Theta_{a}}+I}\right) is not analytically invertible for complex operators such as SAR echo data, we need to identify a specific gradient algorithm to handle complex forward models.

In contrast to conventional proximal gradient (PG) algorithms [49, 50, 51] which alternate between steepest descent and CNN blocks, this paper chooses the CG algorithm to avoid a high total number of iterations. In comparison, CG blocks provide a faster reduction per iteration by enforcing a more accurate data-consistency constraint at each unfolding step. Without trainable parameters, numerous CG steps can be performed at each unfolding with no memory cost during training. Experiments in [38] demonstrate that the CG strategy offers improved performance compared to PG and also facilitates the easy incorporation of other image regularization techniques. In this paper, we set the data consistency block CΘaC_{\Theta_{a}} to 10 iterations using the CG algorithm.

Regarding the CNN block DΦD_{\Phi}, we implement a U-Net model [39], which involves four pooling and unpooling layers with 3×\times3 trainable filters. The parameters of the blocks DΦD_{\Phi} and CΘaC_{\Theta_{a}} are optimized to minimize (21). We rely on the automatic differentiation capability of TensorFlow to evaluate the gradient of the cost function with respect to Θa\Theta_{a} and Φ\Phi.

3.4 Continuous Optimization Training Strategy

In this subsection, we discuss the loss function and backpropagation gradient calculation of the proposed method, which ensure that the optimization training strategy is continuous and differentiable.

3.4.1 Loss Function

In this paper, we specify the loss function as the mean square error between σ^i{Θa,Φ}{\hat{\sigma}}_{i}\left\{{\Theta_{a},\Phi}\right\} and the number of iterations as KK:

{Θa,Φ}=argminΘa,Φi=1NMΘa,Φ(HΘa(si))σi22\begin{split}\left\{{\Theta_{a}^{*},\Phi^{*}}\right\}={\arg{\min\limits_{\Theta_{a},\Phi}{\sum\limits_{i=1}^{N}\left\|{M_{\Theta_{a},\Phi}\left(H_{\Theta_{a}}^{*}\left(s_{i}\right)\right)-\sigma_{i}}\right\|_{2}^{2}}}}\end{split} (30)

where σi\sigma_{i} is the ithi^{th} target image.

3.4.2 Backpropagation and Gradient Calculation

The goal of training is to determine the weight parameter ww, which is shared across the iterations. The gradient of the cost function with respect to the shared weights can be determined using the chain rule

w=k=0K1Jw(zk)T(zk)\begin{split}{\nabla_{w}\mathcal{L}}={\sum\limits_{k=0}^{K-1}{J_{w}\left(z_{k}\right)^{T}\left({\nabla_{z_{k}}\mathcal{L}}\right)}}\end{split} (31)

where the Jacobian matrix Jw(z)J_{w}(z) has entries [Jw(z)]i,j=zi/ωj{\left[{{J_{w}}\left({\rm{z}}\right)}\right]_{i,j}}=\partial{z_{i}}/\partial{\omega_{j}} and zk{z_{k}} is the output of the CNN at the kth{k^{th}} iteration. We now focus on how to backpropagate through the conjugate gradient (CG) blocks. It is important to note that the CG block does not have any trainable parameters. We have

zk1=Jzk1(σk)Tσk\begin{split}{\nabla_{{z_{k-1}}}}{\cal L}={J_{{z_{k-1}}}}{\left({{\sigma_{k}}}\right)^{T}}{\nabla_{{\sigma_{k}}}}{\cal L}\end{split} (32)

where the Jacobian matrix Jz(σ){J_{z}}\left(\sigma\right) has entries [Jz(σ)]i,j=σi/zj{\left[{{J_{z}}\left(\sigma\right)}\right]_{i,j}}=\partial{\sigma_{i}}/\partial{z_{j}}. The Jacobian matrix is given by

Jz(σ)=(HΘaHΘaH+λI)1\begin{split}{J_{z}}\left(\sigma\right)={\left({{H_{{{\rm{\Theta}}_{a}}}}H_{{{\rm{\Theta}}_{a}}}^{H}+\lambda I}\right)^{-1}}\end{split} (33)

Since the Jacobian matrix is symmetric, we have

zk1=(HΘaHΘa+λI)1(σk)\begin{split}{{\nabla_{{z_{k-1}}}}{\cal L}}={\left({{H_{{{\rm{\Theta}}_{a}}}}H_{{{\rm{\Theta}}_{a}}}^{*}+\lambda I}\right)^{-1}}\left({{\nabla_{{\sigma_{k}}}}{\cal L}}\right)\end{split} (34)

We can evaluate the above expression using a CG algorithm, running until convergence, determined by the saturation of the data consistency cost. It should be noted that the above gradient calculation is only valid if we implement (HΘaHΘa+λI)1{\left({{H_{{{\rm{\Theta}}a}}}H{{{\rm{\Theta}}_{a}}}^{*}+\lambda I}\right)^{-1}} or, equivalently, let the CG algorithm converge. This result demonstrates that the gradients can be backpropagated through the CG block. We rely on the variable sharing strategies in TensorFlow to implement the unrolled architecture. The parameters of the networks at each iteration are initialized and updated together.

4 Experiments and Discussions

Refer to caption
Refer to caption
Figure 5: (a) represents oiriginal SAR amplitude image. (b) represents reconstruction SAR amplitude image.

4.1 Experiment Settings

4.1.1 Dataset

In the following experiments, the training and testing data are derived from SSDD sea surface ships SAR images. We use a total of 1300 SAR patches, with sizes standardized to 256×\times256 pixels. We categorize the images into three groups based on scene sparsity: off-shore ships, coastal ships, and strait. The distribution of training and testing datasets for each category is shown in Table 1.

Table 1: Configuration of Training and Testing Dataset
Training Numbers Testing Numbers
Off-shore Ships 600 100
Coastal Ships 200 50
Strait 200 50

Similar to [35], we generate unlabeled training samples by downsampling the original echo data, adding system noise, and introducing echo phase disturbances. This method not only reduces the amount of data and increases imaging efficiency, but also improves the robustness and reliability of the algorithm. The parameters used in these processes are provided in Table 2.

Based on the parameters in Table 2, the amplitude map of the SSDD dataset is used to generate echoes as follows: let the amplitude image of the SSDD dataset be AA, the image after adding a random phase is ASLC{A_{SLC}}, and the point target echo Sp{S_{p}} generated in Table 2. Then, the original echo of the ii-th image Si=SpASLC{S_{i}}={S_{p}}{\rm{*}}{A_{SLC}}. Si{S_{i}} is compared with the original image after imaging as shown in Fig. 5.

Table 2: MAIN PARAMETERS OF SAR SYSTEMS
Parameter Value
Range bandwidth 150 MHz
Range sampling rate 200 MHz
Pluse duration 1 μ\mus
Range chirp rate 200e12 Hz/s
Doppler rate 256.1772 Hz/s
Pulse repetition frequency 200 Hz
Carrier frequency 9.6 GHz
Wavelength 0.0375 m
Azimuth antenna length 2 m
Platform height 10000 m
Equivalent velocity 200 m/s

4.1.2 Evaluation Index Definitions

Structural similarity (SSIM) and peak signal-to-noise ratio (PSNR) are used to quantitatively evaluate the experiment results. Let σ\mathbf{\sigma} be the result of target imaging and σ^\mathbf{\hat{\sigma}} be the target label image. The definitions of these evaluation indexes are as follows:

SSIM(σ^,σ)=(2μσ^μσ+C1)(2ρσ^ρσ+C2)(μσ^2μσ2+C1)(ρσ^2+ρσ2+C2)\begin{split}SSIM\left({\hat{\sigma},\sigma}\right)=\frac{{\left({2{\mu_{\hat{\sigma}}}{\mu_{\sigma}}+{C_{1}}}\right)\left({2{\rho_{\hat{\sigma}}}{\rho_{\sigma}}+{C_{2}}}\right)}}{{\left({\mu_{\hat{\sigma}}^{2}\mu_{\sigma}^{2}+{C_{1}}}\right)\left({\rho_{\hat{\sigma}}^{2}+\rho_{\sigma}^{2}+{C_{2}}}\right)}}\end{split} (35)

where μσ^,μσ,ρσ^,ρσ{\mu_{\hat{\sigma}}},{\mu_{\sigma}},{\rho_{\hat{\sigma}}},{\rho_{\sigma}}, and ρσ^ρσ{\rho_{\hat{\sigma}}}{\rho_{\sigma}} are the local means, standard deviations, cross covariance for images σ^\mathbf{\hat{\sigma}} and σ\mathbf{\sigma}, respectively. C1=(K1L)2{C_{1}}={\left({{K_{1}}L}\right)^{2}}, LL is the variation range of the pixel value, as for 8-bit grayscale image, L=255L=255, K11{K_{1}}\ll 1, C2=(K2L)2{C_{2}}={\left({{K_{2}}L}\right)^{2}}, ow we set K1=0.01K_{1}=0.01 and K2=0.03K_{2}=0.03

PSNR=10log10maxσ22σ^σ22\begin{split}PSNR=10{\log_{10}}\frac{{{\rm{max}}\sigma_{2}^{2}}}{{\hat{\sigma}-\sigma_{2}^{2}}}\end{split} (36)

4.2 Results and Analysis

To demonstrate the performance of image reconstruction and enhancement provided by MF-JMoDL-Net, experiments are conducted under various conditions, such as different sparsity scenes, different sampling rates, and different sampling patterns. Additionally, the azimuth ambiguity suppression capability of the proposed method is compared with the local azimuth ambiguity-to-signal ratio estimation (LAASR) [4] ambiguity suppression proposed by Long. The design of the azimuth sampling pattern using the proposed method is compared with staggered [10, 11, 12, 14, 13], uniform sampling, and Poisson-disk sampling methods [30].

Refer to caption
Figure 6: The performance in terms of azimuth ambiguity suppression under 50%\%, 25%\% and 12.5%\% azimuth undersampling rates.

4.2.1 Results Under Different Azimuth Undersampling Rates

Table 3: results of different undersampling rates and scenes
sampling
rate(%\%)
Scence SSIM(%\%)
Undersamping
PSNR (dB)
Reconstruction
PSNR(dB)
Increase
PSNR (dB)
50 Off-shore ships 0.932306 25.51 37.50 11.99
25 Off-shore ships 0.845824 21.50 33.36 11.86
12.5 Off-shore ships 0.777514 18.25 30.27 12.02
50 Coastal ships 0.724638 14.01 18.69 4.68
50 Strait 0.699019 13.63 18.04 4.41

We use data with 50%, 25%, and 12.5% azimuth undersampling rates to test the influence of the undersampling rate on both the LAASR method and the proposed method. The learned mask, which represents sampling locations, is fixed with the same pattern obtained by the proposed MF-JMoDL-Net. The performance in terms of azimuth ambiguity suppression is shown in Fig. 6.

In Fig. 6, the red boxes indicate locations with severe ambiguity, while the blue boxes show the original target structure. In the case of a 50% undersampling rate, the method proposed in this paper demonstrates better effectiveness in azimuth ambiguity suppression. It is evident that our method’s performance deteriorates as the sampling rate decreases. However, the LAASR method is not affected by undersampling rates and eliminates true targets. The LAASR method relies on the AAP, which fixes the locations of the ambiguity region and removes them awkwardly. In other words, the LAASR method may misjudge targets as ambiguity and eliminate all or part of them. In contrast, the proposed method preserves the target structures more intelligently.

Table 4: results of different undersampling pattern
sampling
rate(%\%)
SSIM(%\%)
Undersamping
PSNR (dB)
Reconstruction
PSNR(dB)
Increase
PSNR(dB)
Proposed 0.932306 25.51 37.50 11.99
Poisson-disk 0.760917 20.67 26.04 5.28
Staggered 0.747645 21.63 25.06 3.43
Refer to caption
(a) Results of off-shore ships under 50%\% undersampling rate.
Refer to caption
(b) Results of coastal ships under 50%\% undersampling rate.
Refer to caption
(c) Results of strait under 50%\% undersampling rate.
Figure 7: The azimuth ambiguity suppression and image reconstruction results of proposed method and LAASR under 50%\% sampling in different scenes.

4.2.2 Results Under Different Scenes

Fig.7 presents the azimuth ambiguity suppression and image reconstruction results of the proposed method and LAASR under a 50% sampling rate for off-shore ships scenes in Fig.7(a), coastal ships scenes in Fig.7(b), and strait scenes in Fig.7(c), respectively. As the scene sparsity decreases, the ambiguity suppression performance of the proposed method remains stable. However, the LAASR method not only fails to suppress ambiguity effectively, but also significantly deteriorates the structural similarity of the original images.

To visually display the performance trends of the proposed method, the curves of SSIM, undersampling PSNR, reconstruction PSNR, and increased PSNR values under different undersampling rates are presented in Table 3. From the experimental results in Table 3, the PSNR remains stable as the sampling rate decreases, even though SSIM declines with the sampling rate.

Refer to caption
Figure 8: Trained sampling intervals under different sparsity scenes.

Fig.8 displays the distribution of distances between sampling positions trained by the proposed MF-JMoDL-Net under typical representative scenes with three degrees of sparsity. In Fig.8, pink dots represent the off-shore ship scenes, blue squares represent coastal scenes, and green diamonds represent strait scenes. Clearly, the intervals between sampling positions change gradually as the scene sparsity decreases. Consequently, it is also worth discussing and verifying whether the optimal sampling pattern approaches uniform sampling when the scene sparsity reaches its lowest level.

4.2.3 Results Under Different Sampling Pattern

We utilize the off-shore ship dataset in Table 1 as the unified training target, and compare the reconstruction results of the proposed method with the staggered sampling pattern and Poisson-disk sampling pattern on the azimuth ambiguity image, as shown in Table 4.

From the results in Table 4 and Fig. 9, the sample interval is shown in the last row. Clearly, the method proposed in this paper performs best in terms of SSIM and PSNR improvement.

Refer to caption
Figure 9: Results of different undersampling patterns.

5 Conclusion

In this paper, we present a joint optimization framework for sparse strip SAR imaging algorithms and azimuth undersampling patterns, namely MF-JMoDL-Net. The architecture of the MF-JMoDL-Net is constructed by introducing the MF approximate measurement operator and inverse MF operator into the JMoDL-based deep network, which serves as the solution for the proposed MF-JMoDL-based imaging model. Through end-to-end training, the extraction and utilization of echo data information are realized, enabling the acquisition of optimal azimuth sampling patterns and the suppression of corresponding ambiguity caused by azimuth undersampling. To verify the proposed method, we conducted experiments and comparisons under various conditions, including different sampling rates (50%, 25%, and 12.5%), different sparsity levels of scenes, and sampling patterns (staggered and Poisson-disk). All of these experiments and comparisons demonstrate the effectiveness and superiority of the proposed MF-JMoDL-Net compared to existing methods.

References

  • [1] J. C. Curlander, R. N. McDonough, Synthetic aperture radar, Vol. 11, Wiley, New York, 1991.
  • [2] A. Moreira, Suppressing the azimuth ambiguities in synthetic aperture radar images, IEEE Transactions on Geoscience and Remote Sensing 31 (4) (1993) 885–895. doi:10.1109/36.239912.
  • [3] J. Chen, M. Iqbal, W. Yang, P.-b. Wang, B. Sun, Mitigation of azimuth ambiguities in spaceborne stripmap sar images using selective restoration, IEEE Transactions on Geoscience and Remote Sensing 52 (7) (2014) 4038–4045. doi:10.1109/TGRS.2013.2279109.
  • [4] Y. Long, F. Zhao, M. Zheng, G. Jin, H. Zhang, An azimuth ambiguity suppression method based on local azimuth ambiguity-to-signal ratio estimation, IEEE Geoscience and Remote Sensing Letters 17 (12) (2020) 2075–2079. doi:10.1109/LGRS.2019.2963126.
  • [5] X. Zhang, Stochastic sampling for antialiasing in computer graphics, Stanford University, 2005.
  • [6] H. S. Shapiro, R. A. Silverman, Alias-free sampling of random noise, Journal of the Society for Industrial and Applied Mathematics 8 (2) (1960) 225–248. doi:10.1137/0108013.
  • [7] E. Masry, Alias-free sampling: An alternative conceptualization and its applications, IEEE Transactions on Information Theory 24 (3) (1978) 317–324. doi:10.1109/TIT.1978.1055889.
  • [8] M. A. Dippé, E. H. Wold, Antialiasing through stochastic sampling, in: Proceedings of the 12th annual conference on Computer graphics and interactive techniques, 1985, pp. 69–78.
  • [9] R. L. Cook, Stochastic sampling in computer graphics, ACM Transactions on Graphics 5 (1) (1986) 51–72. doi:10.1145/7529.8927.
  • [10] M. Villano, G. Krieger, A. Moreira, Staggered-sar for high-resolution wide-swath imaging (2012) 23–23doi:10.1049/cp.2012.1600.
  • [11] M. Villano, G. Krieger, A. Moreira, A novel processing strategy for staggered sar, IEEE Geoscience and Remote Sensing Letters 11 (11) (2014) 1891–1895. doi:10.1109/LGRS.2014.2313138.
  • [12] M. Villano, G. Krieger, A. Moreira, Staggered sar: High-resolution wide-swath imaging by continuous pri variation, IEEE Transactions on Geoscience and Remote Sensing 52 (7) (2014) 4462–4479. doi:10.1109/TGRS.2013.2282192.
  • [13] M. Villano, G. Krieger, M. Jäger, A. Moreira, Staggered sar: Performance analysis and experiments with real data, IEEE Transactions on Geoscience and Remote Sensing 55 (11) (2017) 6617–6638. doi:10.1109/TGRS.2017.2731047.
  • [14] M. Villano, G. Krieger, A. Moreira, Ambiguities and image quality in staggered sar, in: 2015 IEEE 5th Asia-Pacific Conference on Synthetic Aperture Radar (APSAR), 2015, pp. 204–209. doi:10.1109/APSAR.2015.7306189.
  • [15] E. Candes, J. Romberg, T. Tao, Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information, IEEE Transactions on Information Theory 52 (2) (2006) 489–509. doi:10.1109/TIT.2005.862083.
  • [16] D. Donoho, Compressed sensing, IEEE Transactions on Information Theory 52 (4) (2006) 1289–1306. doi:10.1109/TIT.2006.871582.
  • [17] R. Baraniuk, P. Steeghs, Compressive radar imaging, in: 2007 IEEE Radar Conference, 2007, pp. 128–133. doi:10.1109/RADAR.2007.374203.
  • [18] G. Rilling, M. Davies, B. Mulgrew, Compressed sensing based compression of sar raw data, in: SPARS’09-Signal Processing with Adaptive Sparse Structured Representations, 2009.
  • [19] B. Zhang, W. Hong, Y. Wu, Sparse microwave imaging: Principles and applications, Science China Information Sciences 55 (8) (2012) 1722–1754. doi:10.1007/s11432-012-4633-4.
  • [20] C. He, L. Liu, L. Xu, M. Liu, M. Liao, Learning based compressed sensing for sar image super-resolution, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 5 (4) (2012) 1272–1281. doi:10.1109/JSTARS.2012.2189555.
  • [21] M. Cetin, I. Stojanovic, O. Onhon, K. Varshney, S. Samadi, W. C. Karl, A. S. Willsky, Sparsity-driven synthetic aperture radar imaging: Reconstruction, autofocusing, moving targets, and compressed sensing, IEEE Signal Processing Magazine 31 (4) (2014) 27–40. doi:10.1109/MSP.2014.2312834.
  • [22] J. A. Tropp, A. C. Gilbert, Signal recovery from random measurements via orthogonal matching pursuit, IEEE Transactions on Information Theory 53 (12) (2007) 4655–4666. doi:10.1109/TIT.2007.909108.
  • [23] D. Needell, J. Tropp, Cosamp: Iterative signal recovery from incomplete and inaccurate samples, Applied and Computational Harmonic Analysis 26 (3) (2009) 301–321. doi:10.1016/j.acha.2008.07.002.
  • [24] D. Needell, J. A. Tropp, Cosamp: Iterative signal recovery from incomplete and inaccurate samples, Communications of the ACM 53 (12) (2010) 93–100. doi:10.1145/1859204.1859229.
  • [25] A. Beck, M. Teboulle, A fast iterative shrinkage-thresholding algorithm for linear inverse problems, SIAM Journal on Imaging Sciences 2 (1) (2009) 183–202. doi:10.1137/080716542.
  • [26] I. Daubechies, M. Defrise, C. De Mol, An iterative thresholding algorithm for linear inverse problems with a sparsity constraint, Communications on Pure and Applied Mathematics 57 (11) (2004) 1413–1457. doi:10.1002/cpa.20042.
  • [27] J. Fang, Z. Xu, B. Zhang, W. Hong, Y. Wu, Fast compressed sensing sar imaging based on approximated observation, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 7 (1) (2014) 352–363. doi:10.1109/JSTARS.2013.2263309.
  • [28] C. Jiang, B. Zhang, J. Fang, Z. Zhe, W. Hong, Y. Wu, Z. Xu, Efficient lq regularisation algorithm with range–azimuth decoupled for sar imaging, Electronics Letters 50 (3) (2014) 204–205. doi:10.1049/el.2013.1989.
  • [29] J. Sun, Y. Zhang, J. Tian, J. Wang, A novel spaceborne sar wide-swath imaging approach based on poisson disk-like nonuniform sampling and compressive sensing, Science China Information Sciences 55 (8) (2012) 1876–1887. doi:10.1007/s11432-012-4590-y.
  • [30] X. Yang, G. Li, J. Sun, Y. Liu, X.-G. Xia, Sparsity-driven high-resolution and wide-swath sar imaging via poisson disk sampling, in: 2019 IEEE Radar Conference (RadarConf), 2019, pp. 1–5. doi:10.1109/RADAR.2019.8835553.
  • [31] X. Yang, G. Li, J. Sun, Y. Liu, X.-G. Xia, High-resolution and wide-swath sar imaging via poisson disk sampling and iterative shrinkage thresholding, IEEE Transactions on Geoscience and Remote Sensing 57 (7) (2019) 4692–4704. doi:10.1109/TGRS.2019.2892471.
  • [32] V. Monga, Y. Li, Y. C. Eldar, Algorithm unrolling: Interpretable, efficient deep learning for signal and image processing, IEEE Signal Processing Magazine 38 (2) (2021) 18–44. doi:10.1109/MSP.2020.3016905.
  • [33] E. Mason, B. Yonel, B. Yazici, Deep learning for sar image formation, in: E. Zelnio, F. D. Garber (Eds.), SPIE Defense + Security, Anaheim, California, United States, 2017, p. 1020104. doi:10.1117/12.2267831.
  • [34] W. Pu, Deep sar imaging and motion compensation, IEEE Transactions on Image Processing 30 (2021) 2232–2247. doi:10.1109/TIP.2021.3051484.
  • [35] S. Zhao, J. Ni, J. Liang, S. Xiong, Y. Luo, End-to-end sar deep learning imaging method based on sparse optimization, Remote Sensing 13 (21) (2021) 4429. doi:10.3390/rs13214429.
  • [36] M. Li, J. Wu, W. Huo, R. Jiang, Z. Li, J. Yang, H. Li, Target-oriented sar imaging for scr improvement via deep mf-admm-net, IEEE Transactions on Geoscience and Remote Sensing 60 (2022) 1–14. doi:10.1109/TGRS.2022.3145568.
  • [37] H. K. Aggarwal, M. Jacob, J-modl: Joint model-based deep learning for optimized sampling and reconstruction, IEEE Journal of Selected Topics in Signal Processing 14 (6) (2020) 1151–1162. doi:10.1109/JSTSP.2020.3004094.
  • [38] H. K. Aggarwal, M. P. Mani, M. Jacob, Modl: Model-based deep learning architecture for inverse problems, IEEE Transactions on Medical Imaging 38 (2) (2019) 394–405. doi:10.1109/TMI.2018.2865356.
  • [39] O. Ronneberger, P. Fischer, T. Brox, U-net: Convolutional networks for biomedical image segmentation (May 2015). arXiv:1505.04597.
  • [40] G. Ongie, M. Jacob, Super-resolution mri using finite rate of innovation curves, in: 2015 IEEE 12th International Symposium on Biomedical Imaging (ISBI), 2015, pp. 1248–1251. doi:10.1109/ISBI.2015.7164100.
  • [41] G. Ongie, S. Biswas, M. Jacob, Convex recovery of continuous domain piecewise constant images from nonuniform fourier samples, IEEE Transactions on Signal Processing 66 (1) (2018) 236–250. doi:10.1109/TSP.2017.2750111.
  • [42] S. Ma, W. Yin, Y. Zhang, A. Chakraborty, An efficient algorithm for compressed mr imaging using total variation and wavelets, in: 2008 IEEE Conference on Computer Vision and Pattern Recognition, 2008, pp. 1–8. doi:10.1109/CVPR.2008.4587391.
  • [43] M. Figueiredo, R. Nowak, An em algorithm for wavelet-based image restoration, IEEE Transactions on Image Processing 12 (8) (2003) 906–916. doi:10.1109/TIP.2003.814255.
  • [44] M. Jacob, M. P. Mani, J. C. Ye, Structured low-rank algorithms: Theory, magnetic resonance applications, and links to machine learning, IEEE Signal Processing Magazine 37 (1) (2020) 54–68. doi:10.1109/MSP.2019.2950432.
  • [45] H. Chen, Y. Zhang, M. K. Kalra, F. Lin, Y. Chen, P. Liao, J. Zhou, G. Wang, Low-dose ct with a residual encoder-decoder convolutional neural network, IEEE Transactions on Medical Imaging 36 (12) (2017) 2524–2535. doi:10.1109/TMI.2017.2715284.
  • [46] Y. Han, L. Sunwoo, J. C. Ye, $k$ -space deep learning for accelerated mri, IEEE Transactions on Medical Imaging 39 (2) (2020) 377–386. doi:10.1109/TMI.2019.2927101.
  • [47] F. Sherry, M. Benning, J. C. De los Reyes, M. J. Graves, G. Maierhofer, G. Williams, C.-B. Schönlieb, M. J. Ehrhardt, Learning the sampling pattern for mri, IEEE Transactions on Medical Imaging 39 (12) (2020) 4310–4321. doi:10.1109/TMI.2020.3017353.
  • [48] B. Gözcü, R. K. Mahabadi, Y.-H. Li, E. Ilıcak, T. Çukur, J. Scarlett, V. Cevher, Learning-based compressive mri, IEEE Transactions on Medical Imaging 37 (6) (2018) 1394–1406. doi:10.1109/TMI.2018.2832540.
  • [49] K. Hammernik, T. Klatzer, E. Kobler, M. P. Recht, D. K. Sodickson, T. Pock, F. Knoll, Learning a variational network for reconstruction of accelerated mri data, Magnetic Resonance in Medicine 79 (6) (2018) 3055–3071. doi:10.1002/mrm.26977.
  • [50] M. Mardani, H. Monajemi, V. Papyan, S. Vasanawala, D. Donoho, J. Pauly, Recurrent generative adversarial networks for proximal learning and automated compressive image recovery, arXiv preprint arXiv:1711.10046.
  • [51] P. Putzky, M. Welling, Recurrent inference machines for solving inverse problems, arXiv preprint arXiv:1706.04008.