This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Deep Inertia LpL_{p} Half-quadratic Splitting Unrolling Network for Sparse View CT Reconstruction

Yu Guo, Graduate Student Member, IEEE, Caiying Wu, Yaxin Li, Qiyu Jin, Member, IEEE and Tieyong Zeng This work was supported by National Natural Science Foundation of China (No. 12061052), Natural Science Foundation of Inner Mongolia Autonomous Region (No. 2024LHMS01006, 2024MS01002), Young Talents of Science and Technology in Universities of Inner Mongolia Autonomous Region (No. NJYT22090), Special Funds for Graduate Innovation and Entrepreneurship of Inner Mongolia University (No. 11200-121024), Inner Mongolia University Independent Research Project (No. 2022-ZZ004) and the network information center of Inner Mongolia University. (Corresponding author: Caiying Wu)Y. Guo, C. Wu, Y. Li and Q. Jin are with the School of Mathematical Science, Inner Mongolia University, Hohhot 010020, China. (e-mail: [email protected]; [email protected]). T. Zeng is with Department of Mathematics, The Chinese University of Hong Kong, Satin, Hong Kong.
Abstract

Sparse view computed tomography (CT) reconstruction poses a challenging ill-posed inverse problem, necessitating effective regularization techniques. In this letter, we employ LpL_{p}-norm (0<p<10<p<1) regularization to induce sparsity and introduce inertial steps, leading to the development of the inertial LpL_{p}-norm half-quadratic splitting algorithm. We rigorously prove the convergence of this algorithm. Furthermore, we leverage deep learning to initialize the conjugate gradient method, resulting in a deep unrolling network with theoretical guarantees. Our extensive numerical experiments demonstrate that our proposed algorithm surpasses existing methods, particularly excelling in fewer scanned views and complex noise conditions.

Index Terms:
LpL_{p}-norm, inertial, wavelet, unrolling, Sparse View

I Introduction

Computed tomography (CT) is crucial in modern medicine. However, the ionizing radiation from X-rays has led to interest in low-dose CT. Sparse view CT is an innovative approach that reduces projection data, thereby minimizing radiation and shortening scan times. However, it can cause image degradation, including streak artifacts [1]. Mathematically, the CT image reconstruction problem model can be succinctly described as follows:

f=Au+n,f=Au+n, (1)

where ff is the measured low quality image, AA is the measurement operator, uu is the original high quality image, and nn is the noise. In fact, in incomplete data CT reconstruction, the matrix AA is noninvertible. Hence, the inverse problem (1) is ill-posed.

Classical CT reconstruction techniques, such as Algebraic Reconstruction Techniques (ART) [2] and Filtered Backprojection (FBP) [3, 4], struggle to effectively handle sparse view CT. To address this, numerous algorithms have been proposed, with regularization models gaining widespread use. These models typically take the following form:

minu12Auf2+λR(u),\min_{u}\frac{1}{2}\|Au-f\|^{2}+\lambda R(u), (2)

where R(u)R(u) represents the regularization term, and the weighted parameter λ>0\lambda>0 balances the data fidelity and regularization terms. Various regularization terms, such as BM3D [5, 6, 7], total variation [8], sparsity [9, 10], low rank [11], and wavelets [12, 13, 14], have been employed to tackle this challenge.

With the advent of deep learning in image processing, an increasing number of sparse view CT reconstructions are leveraging deep learning algorithms to achieve superior performance. These primarily include direct reconstruction using end-to-end network structures [15, 16], and deep unrolling combined with traditional algorithms [17, 13, 18, 19, 20]. The advantage of deep unfolding networks is that they are driven by both model and data [21], meaning that prior knowledge in the domain can assist in learning. Recently, there are new techniques introduced to deep learning, such as attention mechanisms [22], and diffusion models [23, 24, 25], etc. Currently, state-of-the-art sparse-view CT reconstruction results are obtained by score-based model [26] and diffusion model [27]. While deep learning algorithms offer improved performance, they are dependent on high-quality pairwise data and lack theoretical guarantees, rendering them as black-box algorithms.

In this letter, we address the sparse view CT reconstruction problem by leveraging the superior sparsity of the LpL_{p}-norm (0<p<10<p<1) in conjunction with wavelet transform. To tackle this non-convex and non-continuous problem, we introduce the Inertial LpL_{p}-norm Half-Quadratic Splitting method (IHQSp), which employs inertial steps to expedite convergence. We establish the convergence of the IHQSp algorithm. During the CT image reconstruction, the corresponding subproblem is resolved using the conjugate gradient (CG) method. Considering the n-step quadratic convergence of CG, we propose using a network trained via deep learning as the initializer for CG. In the context of the IHQSp algorithm, the network solely provides the initial value for the CG, thereby not influencing the algorithm’s convergence. From a deep learning perspective, we derive a learning algorithm IHQSp-Net with convergence guarantees, which is deep unrolled by the IHQSp algorithm. We validate our proposed IHQSp-CG and IHQSp-Net through extensive numerical experiments on two datasets, demonstrating their exceptional performance.

II Inertial HQSp-CG Algorithm and Convergence Analysis

Algorithm 1 Inertial HQSp-CG algorithm
  Initialization: u0=uFBPu^{0}=u_{{\rm FBP}}, z0=Wu0,u¯0=u0,z¯0=z0,α,β,λ,γ,ϵz^{0}=Wu^{0},\bar{u}^{0}=u^{0},\bar{z}^{0}=z^{0},\alpha,\beta,\lambda,\gamma,\epsilon.
  for k=0:Maxlterk=0:\mathrm{Maxlter} do:
    1. Updata zk+1=proxp,γλ(Wu¯k)z^{k+1}={\rm prox}_{p,\frac{\gamma}{\lambda}}(W\bar{u}^{k}).
    2. Updata z¯k+1=zk+1+α(zk+1z¯k)\bar{z}^{k+1}=z^{k+1}+\alpha(z^{k+1}-\bar{z}^{k}).
    3. Updata uk+1=CG(uk,z¯k+1,γ)u^{k+1}={\rm CG}(u^{k},\bar{z}^{k+1},\gamma).
    4. Updata u¯k+1=uk+1+β(uk+1u¯k)\bar{u}^{k+1}=u^{k+1}+\beta(u^{k+1}-\bar{u}^{k}).
    5. Stopping criterion: u¯k+1u¯ku¯kϵ\frac{\|\bar{u}^{k+1}-\bar{u}^{k}\|}{\|\bar{u}^{k}\|}\leq\epsilon.
  end.
  Output: u¯k+1\bar{u}^{k+1}

II-A Inertial HQSp-CG Algorithm

In order to take advantage of the sparsity of the CT image, we use the LpL_{p}-norm with the wavelet transform to form the regularization term in (2), i.e.

minu12Auf2+λWupp,\min_{u}\frac{1}{2}\|Au-f\|^{2}+\lambda\|Wu\|^{p}_{p}, (3)

where 0<p<10<p<1, W=(W1,W2,,Wm)W=(W_{1},W_{2},...,W_{m}) is a mm-channel operator with WiTWi=I(1im)W_{i}^{T}W_{i}=I(1\leq i\leq m). The operator WW is chosen as the highpass components of the piecewise linear tight wavelet frame transform [28]. Solving (3) directly is challenging due to the discontinuity in the LpL_{p} norm. To solve (3), we introduce the auxiliary variable zz, and use the half-quadratic splitting method to obtain the following problem:

minu,z12Auf22+λzpp+12i=1mγiWiuzi22,\min_{u,z}\frac{1}{2}\|Au-f\|^{2}_{2}+\lambda\|z\|^{p}_{p}+\frac{1}{2}\sum_{i=1}^{m}\gamma_{i}\|W_{i}u-z_{i}\|^{2}_{2}, (4)

where λ>0\lambda>0 and γ=(γ1,γ2,,γm)\gamma=(\gamma_{1},\gamma_{2},...,\gamma_{m}) with γi>0\gamma_{i}>0 (i=1,2,,m)(i=1,2,...,m) are weight parameters. The above problem can be split into zz-subproblem and uu-subproblem solved in alternating iterations. To speed up the convergence, inspired by the inertia algorithm [29, 30], we introduce inertia steps in solving the two subproblems, i.e.,

zk+1\displaystyle z^{k+1} =argminzλzpp+12i=1mγiWiu¯kzi22,\displaystyle=\arg\min_{z}\lambda\|z\|^{p}_{p}+\frac{1}{2}\sum_{i=1}^{m}\gamma_{i}\|W_{i}\bar{u}^{k}-z_{i}\|^{2}_{2}, (5)
z¯k+1\displaystyle\bar{z}^{k+1} =zk+1+α(zk+1z¯k),\displaystyle=z^{k+1}+\alpha(z^{k+1}-\bar{z}^{k}), (6)
uk+1\displaystyle u^{k+1} =argminu12Auf22+12i=1mγiWiuz¯ik+122,\displaystyle=\arg\min_{u}\frac{1}{2}\|Au-f\|^{2}_{2}+\frac{1}{2}\sum_{i=1}^{m}\gamma_{i}\|W_{i}u-\bar{z}_{i}^{k+1}\|^{2}_{2}, (7)
u¯k+1\displaystyle\bar{u}^{k+1} =uk+1+β(uk+1u¯k),\displaystyle=u^{k+1}+\beta(u^{k+1}-\bar{u}^{k}), (8)

where (5) and (7) are subproblem solving steps, (6) and (8) are inertia steps, and α[0,1)\alpha\in[0,1) and β[0,512)\beta\in[0,\frac{\sqrt{5}-1}{2}) are inertia weight parameters.

The zk+1z^{k+1}-update step (5) can be solved by the proximal operator proxp,η{\rm prox}_{p,\eta}, which is defined as

proxp,η(t)=argminxxpp+η2xt2.{\rm prox}_{p,\eta}(t)=\mathop{\arg\min}\limits_{x}{\|x\|_{p}^{p}+\frac{\eta}{2}\|x-t\|^{2}}. (9)

where 0<p<10<p<1, it can be computed as [31]

proxp,η(t)={0,|t|<τ{0,sign(t)ρ},|t|=τsign(t)g,|t|>τ{\rm prox}_{p,\eta}(t)=\left\{\begin{array}[]{ll}0,&\left|t\right|<\tau\\ \{0,{\rm sign}(t)\rho\},&\left|t\right|=\tau\\ {\rm sign}(t)g,&\left|t\right|>\tau\end{array}\right. (10)

where ρ=(2(1p)/η)12p\rho=(2(1-p)/\eta)^{\frac{1}{2-p}}, τ=ρ+pρp1/η\tau=\rho+p\rho^{p-1}/\eta, gg is the solution of h(g)=pgp1+ηgη|t|=0h(g)=pg^{p-1}+\eta g-\eta|t|=0 over the region (ρ,|t|)(\rho,|t|). Since h(g)h(g) is convex, when |t|>τ|t|>\tau, gg can be efficiently solved using Newton’s method.

For the uk+1u^{k+1}-update step (7), according to the Euler-Lagrange equations, we have:

uk+1=(ATA+i=1mγiWiTWi)1(ATf+i=1mγiWiTz¯ik).\displaystyle u^{k+1}=(A^{T}A+\sum_{i=1}^{m}\gamma_{i}W_{i}^{T}W_{i})^{-1}(A^{T}f+\sum_{i=1}^{m}\gamma_{i}W_{i}^{T}\bar{z}_{i}^{k}).

To avoid the inverse, this equation can be solved using the conjugate gradient method. We call our algorithm as Inertial HQSp-CG (IHQSp-CG) for short and summarized in Algorithm 1. Next, we establish the global convergence of the IHQSp-CG algorithm.

Refer to caption
Refer to caption
Figure 1: IHQSp-Net algorithm framework diagram.

II-B Convergence analysis

In this section, we provide a convergence analysis of Algorithm 1.

Lemma II.1.

Suppose that the sequences {vk}\left\{v^{k}\right\} and {v¯k}\left\{\bar{v}^{k}\right\} generated via Algorithm 1, 0<β<5120<\beta<\frac{\sqrt{5}-1}{2}, then the sets {vk}\left\{v^{k}\right\} and {v¯k}\left\{\bar{v}^{k}\right\} are bounded.

Theorem II.1.

Let {vk}\left\{v^{k}\right\} and {v¯k}\left\{\bar{v}^{k}\right\} be the sequences generated by our algorithm. Then any cluster point of {vk}\left\{v^{k}\right\} is the global minimum point of LL.

Details of the proof can be found in the supplementary material.

TABLE I: Comparison of results between different algorithms on the AAPM and Covid-19 datasets (PSNR/SSIM). The best value of the traditional algorithm is marked with bold. The best value for deep learning is marked with red.
Data AAPM Dataset
Method Traditional algorithm Deep learning algorithm
Method FBP BM3D HQS-CG HQSp-CG PWLS-CSCGR IHQSp-CG PDNet FBPNet MetaInvNet IHQSp-Net* IHQSp-Net
Noise σ=0.3\sigma=0.3
180 23.64/0.5136 24.82/0.6839 31.65/0.8047 32.10/0.8131 26.72/24.7975 32.16/0.8138 30.06/0.8124 31.73/0.8138 33.01/0.8348 32.70/0.8297 33.18/ 0.8415
120 20.69/0.4104 23.31/0.6454 30.11/0.7659 30.72/0.7841 26.97/23.9243 30.86/0.7865 28.89/0.7813 30.36/0.7811 32.17/0.8103 31.45/0.7999 32.08/ 0.8148
90 20.11/0.3494 23.55/0.6143 28.83/0.7324 29.59/0.7602 25.76/22.8060 29.81/0.7647 28.26/0.7633 29.60/0.7674 31.56/0.7939 30.47/0.7777 31.66/ 0.8032
60 18.21/0.2762 22.34/0.5580 26.84/0.6809 27.79/0.7226 23.94/21.0367 28.06/0.7298 26.61/0.7298 28.06/0.7397 30.38/0.7627 28.89/0.7399 30.24/ 0.7671
Noise σ=0.3\sigma=0.3 and I0=5×105I_{0}=5\times 10^{5}
180 22.77/0.4397 24.82/0.6830 30.57/0.7641 30.91/0.7845 25.98/22.0973 30.92/0.7858 29.54/0.7922 31.38/0.7990 32.02/0.8080 31.53/0.8005 32.25/ 0.8157
120 20.04/0.3494 23.31/0.6443 29.08/0.7179 29.81/0.7595 22.45/14.5781 29.90/0.7619 28.17/0.7633 30.15/0.7750 31.33/0.7929 30.48/0.7743 31.40/ 0.7961
90 19.37/0.2968 23.55/0.6132 27.91/0.6832 28.78/0.7369 22.94/15.7751 28.97/0.7430 27.16/0.7363 29.19/0.7570 30.76/0.7674 29.71/0.7516 30.98/ 0.7815
60 17.55/0.2341 22.37/0.5575 26.18/0.6359 27.09/0.6961 22.63/16.3278 27.30/0.7049 25.51/0.7077 28.13/0.7332 29.66/0.7424 28.62/0.7213 30.08/ 0.7597
Data Covid-19 Dataset
Method FBP BM3D HQS-CG HQSp-CG PWLS-CSCGR IHQSp-CG PDNet FBPNet MetaInvNet IHQSp-Net* IHQSp-Net
Noise σ=0.3\sigma=0.3
180 23.91/0.5361 25.46/0.7304 33.32/0.8393 33.93/0.8496 28.50/20.5285 34.05/0.8515 30.71/0.8413 33.21/0.8512 34.76/0.8638 34.43/0.8584 35.03/ 0.8704
120 21.07/0.4303 24.19/0.6906 31.56/0.8041 32.29/0.8241 28.25/19.7033 32.64/0.8285 29.61/0.8140 31.66/0.8196 34.06/ 0.8523 32.95/0.8290 33.83/0.8489
90 20.50/0.3665 24.11/0.6585 29.91/0.7707 30.83/0.8015 27.01/19.1825 31.29/0.8078 29.12/0.7980 30.81/0.8074 33.25/0.8380 31.93/0.8120 33.52/ 0.8432
60 18.47/0.2897 22.26/0.5966 27.30/0.7179 28.36/0.7621 24.98/17.8874 28.94/0.7731 27.34/0.7651 28.36/0.7740 31.88/ 0.8150 30.12/0.7729 31.81/0.8147
Noise σ=0.3\sigma=0.3 and I0=5×105I_{0}=5\times 10^{5}
180 22.96/0.4426 25.45/0.7294 31.81/0.7919 32.30/0.8163 27.44/17.9762 32.34/0.8188 30.13/0.8199 32.84/0.8367 33.69/0.8452 32.75/0.8250 33.98/ 0.8482
120 20.32/0.3515 24.18/0.6893 30.17/0.7480 31.10/0.7967 22.97/11.6984 31.33/0.8009 29.35/0.7979 31.45/0.8154 32.87/0.8302 31.65/0.7993 33.06/ 0.8326
90 19.65/0.2980 24.10/0.6572 28.78/0.7147 29.81/0.7748 23.63/13.0201 30.21/0.7832 27.78/0.7682 30.45/0.7979 32.18/0.8150 30.79/0.7787 32.63/ 0.8239
60 17.71/0.2340 22.27/0.5956 26.58/0.6674 27.60/0.7328 23.41/13.8117 28.05/0.7444 26.25/0.7414 28.80/0.7729 30.80/0.7897 29.39/0.7493 31.62/ 0.8048
Cpu (s) 0.99 3.45 18.63 34.72 26.24 2.03 5.18 6.76 2.08 6.98
Gpu (s) 8.86 17.63 1530.59 15.91 0.62 1.35 1.92 1.98 2.29

III Conjugate gradient initialization and deep unrolling networks

The algorithm, as per Algorithm 1, is divided into two subproblems: z-subproblem for denoising and u-subproblem for CT image reconstruction. The u-subproblem solution uses the conjugate gradient method, known for its super-linear convergence rate. The convergence speed is tied to the spectral distribution of the coefficient matrix, with faster convergence when eigenvalues are concentrated and the condition number is smaller [32, 33]. A well-defined initial value can lead to finite iterations for convergence. To facilitate this, we introduce a deep learning network to adjust the input u¯k\bar{u}^{k} at each iteration:

u^k=u¯k+Net(u¯k;θ),\hat{u}^{k}=\bar{u}^{k}+{\rm Net}(\bar{u}^{k};\theta),

where θ\theta represents the learnable network parameter, and u^k\hat{u}^{k} is the corrected result. The algorithm’s structure can incorporate any network structure, with UNet and lightweight DnCNN as examples, referred to as IHQSp-Net and IHQSp-Net*, respectively. A detailed description is shown in Fig. 1. The parameter θ\theta is trained through a loss function:

(θ)=1Kk=1K(μ12(u^k,f)+μ2SSIM(u^k,f)),\mathcal{L}(\theta)=\frac{1}{K}\sum_{k=1}^{K}(\mu_{1}\mathcal{L}_{2}(\hat{u}^{k},f)+\mu_{2}\mathcal{L}_{{\rm SSIM}}(\hat{u}^{k},f)),

where μ1\mu_{1} and μ2\mu_{2} are weight factors, we set μ1=μ2=1\mu_{1}=\mu_{2}=1. 2\mathcal{L}_{2} is the L2L_{2} loss, i.e. 2(x,y)=xy2\mathcal{L}_{2}(x,y)=\sum\|x-y\|^{2}. SSIM\mathcal{L}_{SSIM} is the structural similarity loss, i.e. SSIM(x,y)=1SSIM(x,y)\mathcal{L}_{SSIM}(x,y)=1-\mathrm{SSIM}(x,y) [34].

From an algorithmic view, the deep network is embedded in the splitting algorithm, serving as an initializer for CG solving without affecting convergence. From a deep learning view, the splitting algorithm framework is a deep unrolling network with theoretical guarantees. The integration of both enhances performance while ensuring theoretical robustness.

IV Numerical experiments

IV-A Datasets and experimental settings

The deep learning algorithm used training and validation data from the “2016 NIH-AAPM-Mayo Clinic Low Dose CT Grand Challenge” dataset [35]. The training set had 1596 slices from five patients, and the validation set had 287 slices from one patient. Experiments were uniformly performed using a fan beam geometry with 800 detector elements. The test data included 38 slices from new patients in the AAPM dataset and 30 slices from the Covid-19 dataset [36] to assess the algorithm’s generalization and robustness.

Our proposed algorithms were compared with eight others, including five traditional (FBP [3], BM3D [5], HQS-CG [13], combining compressed sensing and sparse convolutional coding for PWLS-CSCGR [10], HQSp-CG without inertial LpL_{p} norms) and three deep learning algorithms (PDNet [17], FBPNet [15], MetaInvNet [13]). For our lightweight IHQSp-Net* and IHQSp-Net, we set K = 6 and limited the CG algorithm iterations to 7. Data were degraded with different angles (60, 90, 120, 180) and two noise levels (Gaussian noise σ=0.3\sigma=0.3, mixed Gaussian and Poisson noise with intensity I0=5×105I_{0}=5\times 10^{5}).

Our models were implemented using the PyTorch framework. The experiments were conducted on PyTorch 1.3.1 backend with an NVIDIA Tesla V100S GPU, 8-core CPU and 32GB RAM. For IHQSp-CG, the hyperparameters were set to p=0.7,α=0.5p=0.7,\alpha=0.5, β=0.6\beta=0.6, λ(0.0004,0.0008)\lambda\in(0.0004,0.0008), γ(0.1,0.4)\gamma\in(0.1,0.4). For IHQSp-Net, the hyperparameters were set to p=0.7,α=0.3p=0.7,\alpha=0.3, β=0.1\beta=0.1, λ(0.001,0.004)\lambda\in(0.001,0.004), γ(0.01,0.06)\gamma\in(0.01,0.06).

Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
(a) Ground truth (b) FBP (c) BM3D (d) HQS-CG (e) HQSp-CG (f) PWLS-CSCGR
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
(g) IHQSp-CG (h) PDNet (i) FBPNet (j) MetaInvNet (k) IHQSp-Net* (l) IHQSp-Net
Figure 2: 120 sparse views CT image reconstruction results and magnified ROIs from Covid-19. The sinogram is corrupted by σ=0.3\sigma=0.3 and I0=5×105I_{0}=5\times 10^{5}.
Refer to caption Refer to caption
Figure 3: Robustness analysis of all compared algorithms.

IV-B Comparison of quantitative and qualitative results

Evaluation metrics used are peak signal-to-noise ratio (PSNR) [37] and structural similarity (SSIM) [34]. For a comprehensive evaluation of the algorithm performance, we added a comparison of the mean absolute error (MAE) and the root mean square error (RMSE) in the supplementary material. Table I shows that our proposed IHQSp-CG algorithm outperforms other traditional algorithms, even surpassing the deep learning algorithms PDNet and FBPNet on both AAPM and Covid-19 datasets. It shows greater robustness to noise and missing range than the HQS-CG algorithm, with an average improvement of more than 0.5 dB for 180-degree and 120-degree observations, and over 1 dB for 90-degree and 60-degree observations.

Our proposed IHQSp-Net outperforms other deep learning methods, especially in handling mixed noises. Our lightweight IHQSp-Net* significantly improves upon the traditional IHQSp-CG, indicating a lightweight CG initializer can enhance performance and convergence speed. Compared to FBPNet, IHQSp-Net shows an average improvement of 2 dB, highlighting the benefits of integrating traditional algorithms with theoretical guarantees into deep networks. Meanwhile, Table I gives the running time for each algorithm to reconstruct a 512×\times512 observation of 90 degrees.

Fig. 2 compares reconstructed images. FBP, BM3D, and HQS-CG show artifacts and missing structures due to lack of observation angle. Our IHQSp-CG yields smoother results. Deep learning algorithms like PDNet and FBPNet produce smooth but over-smoothed results, blurring structural information. Both MetaInvNet and our IHQSp-Net produce pleasing results, but IHQSp-Net captures more details and avoids pseudo-connections seen in MetaInvNet. Notably, IHQSp-Net* visually outperforms IHQSp-CG, highlighting the importance of the CG initializer. The robustness analysis of all the algorithms is given in Fig. 3. By statistically analyzing the reconstructed images for all noise cases, observation angle cases and datasets, it can be found that the proposed IHQSp-CG and IHQSp-Net are robust to noise, observation angle, and data.

Refer to caption Refer to caption
(a) pp (b) Convergence
Figure 4: (a) Regarding the ablation experiment of pp-value. (b) The impact of inertia step on algorithm convergence.

IV-C Ablation experiment

There is a key parameter pp in IHQSp which controls the sparsity of the regular term. Different values of pp affect the reconstruction results of CT images [9]. Therefore, we need to discuss the different pp-value effects. Fig. 4(a) shows the results of different pp values on two datasets, AAPM and Covid-19. From the figure, we can find that between 1 and 0.7, the reconstruction effect is enhanced as pp decreases. This reflects the advantage of the LpL_{p} paradigm over L1L_{1}. However, lowerpp is not better. As can be seen in the figure, for IHQSp, the optimal p=0.7p=0.7.

One of the contributions of IHQSp is the introduction of inertial steps to accelerate the convergence of HQS. HQSp-CG, in Table I, shows the numerical results without inertia step. It can be found that the introduction of the inertia step can effectively improve the algorithm’s performance, and the improvement becomes more obvious as the observation angle becomes sparser. Fig. 4(b) gives a comparison between IHQSp and HQSp-CG for iterative reconstruction when the 90-degree observation is corrupted by mixed Gaussian Poisson noise. PSNR and error ukuk12uk2\frac{||u^{k}-u^{k-1}||^{2}}{||u^{k}||^{2}} are used as metrics. From Fig. 4(b), IHQSp stops iterating at step 65, while HQSp without inertial steps iterates until step 90. Meanwhile, HQSp with more iterations does not outperform IHQSp in PSNR. This reflects the importance of introducing inertial steps.

V Conclusion

We propose an inertial LpL_{p}-norm half-quadratic splitting algorithm for sparse view CT image reconstruction and establish its convergence. Building on IHQSp, we introduce IHQSp-Net, a deep unrolling network that comes with theoretical guarantees. Both quantitative and qualitative comparisons with other methods demonstrate that our proposed IHQSp-CG and IHQSp-Net outperform in terms of performance and robustness, particularly excelling in scenarios with fewer scanned views and complex noise situations.

References

  • [1] J. Bian, J. H. Siewerdsen, X. Han, E. Y. Sidky, J. L. Prince, C. A. Pelizzari, and X. Pan, “Evaluation of sparse-view reconstruction from flat-panel-detector cone-beam ct,” Phys. Med. Biol., vol. 55, no. 22, p. 6575, 2010.
  • [2] R. Gordon, R. Bender, and G. T. Herman, “Algebraic reconstruction techniques (art) for three-dimensional electron microscopy and x-ray photography,” J. Theor. Biol., vol. 29, no. 3, pp. 471–481, 1970.
  • [3] A. Katsevich, “Theoretically exact filtered backprojection-type inversion algorithm for spiral ct,” SIAM J. Appl. Math., vol. 62, no. 6, pp. 2012–2026, 2002.
  • [4] Y. Wei, G. Wang, and J. Hsieh, “Relation between the filtered backprojection algorithm and the backprojection algorithm in ct,” IEEE Signal Process. Lett., vol. 12, no. 9, pp. 633–636, 2005.
  • [5] K. Dabov, A. Foi, V. Katkovnik, and K. Egiazarian, “Image denoising by sparse 3-d transform-domain collaborative filtering,” IEEE Trans. Image Process., vol. 16, no. 8, pp. 2080–2095, 2007.
  • [6] T. Zhao, J. Hoffman, M. McNitt-Gray, and D. Ruan, “Ultra-low-dose ct image denoising using modified bm3d scheme tailored to data statistics,” Med. Phys., vol. 46, no. 1, pp. 190–198, 2019.
  • [7] Y. Guo, A. Davy, G. Facciolo, J.-M. Morel, and Q. Jin, “Fast, nonlocal and neural: a lightweight high quality solution to image denoising,” IEEE Signal Process. Lett., vol. 28, pp. 1515–1519, 2021.
  • [8] F. Mahmood, N. Shahid, U. Skoglund, and P. Vandergheynst, “Adaptive graph-based total variation for tomographic reconstructions,” IEEE Signal Process. Lett., vol. 25, no. 5, pp. 700–704, 2018.
  • [9] L. Zhang, H. Zhao, W. Ma, J. Jiang, L. Zhang, J. Li, F. Gao, and Z. Zhou, “Resolution and noise performance of sparse view x-ray ct reconstruction via lp-norm regularization,” Phys. Medica, vol. 52, pp. 72–80, 2018.
  • [10] P. Bao, W. Xia, K. Yang, W. Chen, M. Chen, Y. Xi, S. Niu, J. Zhou, H. Zhang, H. Sun, Z. Wang, and Y. Zhang, “Convolutional sparse coding for compressed sensing ct reconstruction,” IEEE Trans. Med. Imaging, vol. 38, no. 11, pp. 2607–2619, 2019.
  • [11] K. Kim, J. C. Ye, W. Worstell, J. Ouyang, Y. Rakvongthai, G. El Fakhri, and Q. Li, “Sparse-view spectral ct reconstruction using spectral patch-based low-rank penalty,” IEEE Trans. Med. Imaging, vol. 34, no. 3, pp. 748–760, 2014.
  • [12] B. Dong, J. Li, and Z. Shen, “X-ray ct image reconstruction via wavelet frame based regularization and radon domain inpainting,” J. Sci. Comput., vol. 54, no. 2, pp. 333–349, 2013.
  • [13] H. Zhang, B. Liu, H. Yu, and B. Dong, “Metainv-net: meta inversion network for sparse view ct image reconstruction,” IEEE Trans. Med. Imaging, vol. 40, no. 2, pp. 621–634, 2020.
  • [14] E. Sakhaee and A. Entezari, “Joint inverse problems for signal reconstruction via dictionary splitting,” IEEE Signal Process. Lett., vol. 24, no. 8, pp. 1203–1207, 2017.
  • [15] K. H. Jin, M. T. McCann, E. Froustey, and M. Unser, “Deep convolutional neural network for inverse problems in imaging,” IEEE Trans. Image Process., vol. 26, no. 9, pp. 4509–4522, 2017.
  • [16] W. Du, H. Chen, P. Liao, H. Yang, G. Wang, and Y. Zhang, “Visual attention network for low-dose ct,” IEEE Signal Process. Lett., vol. 26, no. 8, pp. 1152–1156, 2019.
  • [17] J. Adler and O. Öktem, “Learned primal-dual reconstruction,” IEEE Trans. Med. Imaging, vol. 37, no. 6, pp. 1322–1332, 2018.
  • [18] B. Shi, K. Jiang, S. Zhang, Q. Lian, Y. Qin, and Y. Zhao, “Mud-net: multi-domain deep unrolling network for simultaneous sparse-view and metal artifact reduction in computed tomography,” Machine Learning: Science and Technology, vol. 5, no. 1, p. 015010, 2024.
  • [19] J. Pan, H. Yu, Z. Gao, S. Wang, H. Zhang, and W. Wu, “Iterative residual optimization network for limited-angle tomographic reconstruction,” IEEE Trans. Image Process., vol. 33, pp. 910–925, 2024.
  • [20] W. Wu, D. Hu, C. Niu, H. Yu, V. Vardhanabhuti, and G. Wang, “Drone: Dual-domain residual-based optimization network for sparse-view ct reconstruction,” IEEE Trans. Med. Imaging, vol. 40, no. 11, pp. 3002–3014, 2021.
  • [21] B. Shi, S. Zhang, K. Jiang, and Q. Lian, “Coupling model- and data-driven networks for ct metal artifact reduction,” IEEE Trans. Comput. Imaging, vol. 10, pp. 415–428, 2024.
  • [22] W. Wu, X. Guo, Y. Chen, S. Wang, and J. Chen, “Deep embedding-attention-refinement for sparse-view ct reconstruction,” IEEE Trans. Instrum. Meas., vol. 72, pp. 1–11, 2023.
  • [23] W. Wu, Y. Wang, Q. Liu, G. Wang, and J. Zhang, “Wavelet-improved score-based generative model for medical imaging,” IEEE Trans. Med. Imaging, vol. 43, no. 3, pp. 966–979, 2024.
  • [24] K. Xu, S. Lu, B. Huang, W. Wu, and Q. Liu, “Stage-by-stage wavelet optimization refinement diffusion model for sparse-view ct reconstruction,” IEEE Trans. Med. Imaging, pp. 1–1, 2024.
  • [25] W. Wu, J. Pan, Y. Wang, S. Wang, and J. Zhang, “Multi-channel optimization generative model for stable ultra-sparse-view ct reconstruction,” IEEE Trans. Med. Imaging, pp. 1–1, 2024.
  • [26] Y. Wang, Z. Li, and W. Wu, “Time-reversion fast-sampling score-based model for limited-angle ct reconstruction,” IEEE Trans. Med. Imaging, pp. 1–1, 2024.
  • [27] Z. Li, D. Chang, Z. Zhang, F. Luo, Q. Liu, J. Zhang, G. Yang, and W. Wu, “Dual-domain collaborative diffusion sampling for multi-source stationary computed tomography reconstruction,” IEEE Trans. Med. Imaging, pp. 1–1, 2024.
  • [28] A. Ron and Z. Shen, “Affine systems inl2 (rd): the analysis of the analysis operator,” J. Funct. Anal., vol. 148, no. 2, pp. 408–447, 1997.
  • [29] F. Alvarez and H. Attouch, “An inertial proximal method for maximal monotone operators via discretization of a nonlinear oscillator with damping,” Set-Valued Anal., vol. 9, pp. 3–11, 2001.
  • [30] P. Ochs, Y. Chen, T. Brox, and T. Pock, “ipiano: Inertial proximal algorithm for nonconvex optimization,” SIAM J. Imaging Sci., vol. 7, no. 2, pp. 1388–1419, 2014.
  • [31] G. Marjanovic and V. Solo, “On l_ql\_q optimization and matrix completion,” IEEE Trans. Signal Process., vol. 60, no. 11, pp. 5714–5724, 2012.
  • [32] D. Smyl, T. N. Tallman, D. Liu, and A. Hauptmann, “An efficient quasi-newton method for nonlinear inverse problems via learned singular values,” IEEE Signal Process. Lett., vol. 28, pp. 748–752, 2021.
  • [33] M. Savanier, E. Chouzenoux, J.-C. Pesquet, and C. Riddell, “Unmatched preconditioning of the proximal gradient algorithm,” IEEE Signal Process. Lett., vol. 29, pp. 1122–1126, 2022.
  • [34] Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE Trans. Image Process., vol. 13, no. 4, pp. 600–612, 2004.
  • [35] C. McCollough, “Tu-fg-207a-04: overview of the low dose ct grand challenge,” Med. Phys., vol. 43, no. 6Part35, pp. 3759–3760, 2016.
  • [36] Y. Shi, G. Wang, X.-p. Cai, J.-w. Deng, L. Zheng, H.-h. Zhu, M. Zheng, B. Yang, and Z. Chen, “An overview of covid-19,” J. Zhejiang Univ.-SCI. B, vol. 21, no. 5, pp. 343–360, 2020.
  • [37] A. Hore and D. Ziou, “Image quality metrics: Psnr vs. ssim,” in Proc. IEEE Conf. Comput. Vision Pattern Recognit. (CVPR).   IEEE, 2010, pp. 2366–2369.

VI Convergence analysis

Define vk=(zk,uk),v¯k=(z¯k,u¯k)v^{k}=(z^{k},u^{k}),\bar{v}^{k}=(\bar{z}^{k},\bar{u}^{k}),

L(vk)=12Aukf22+λzkpp+12i=1mγiWiukzik22.\displaystyle L(v^{k})=\frac{1}{2}\|Au^{k}-f\|^{2}_{2}+\lambda\|z^{k}\|^{p}_{p}+\frac{1}{2}\sum_{i=1}^{m}\gamma_{i}\|W_{i}u^{k}-z_{i}^{k}\|^{2}_{2}.
Lemma VI.1.

Suppose that the sequences {vk}\left\{v^{k}\right\} and {v¯k}\left\{\bar{v}^{k}\right\} generated via Algorithm 1, 0<β<5120<\beta<\frac{\sqrt{5}-1}{2}, then the sets {vk}\left\{v^{k}\right\} and {v¯k}\left\{\bar{v}^{k}\right\} are bounded.

Proof.

In our problem, uku^{k} represents the gray value of an image, i.e., 0ui,jk2550\leq u^{k}_{i,j}\leq 255, therefore {uk}\left\{u^{k}\right\} is bounded. From the uk+1u^{k+1}-update step (7), we have

uk+1=(ATA+i=1mγiWiTWi)1(ATf+i=1mγiWiTz¯ik),u^{k+1}=(A^{T}A+\sum_{i=1}^{m}\gamma_{i}W_{i}^{T}W_{i})^{-1}(A^{T}f+\sum_{i=1}^{m}\gamma_{i}W_{i}^{T}\bar{z}_{i}^{k}),

which together with WiTWi=IW_{i}^{T}W_{i}=I and WiW_{i} is the given wavelet transform operator, yields,

i=1mγiWiTz¯ik=(ATA+i=1mγiI)uk+1ATf.\sum_{i=1}^{m}\gamma_{i}W_{i}^{T}\bar{z}_{i}^{k}=(A^{T}A+\sum_{i=1}^{m}\gamma_{i}I)u^{k+1}-A^{T}f.

From {uk}\left\{u^{k}\right\} is bounded, it is easy to see that {z¯k}\left\{\bar{z}^{k}\right\} is bounded.

On the other hand, from (8), we find

u¯k+1\displaystyle\bar{u}^{k+1} =uk+1+β(uk+1u¯k)=(1+β)uk+1βu¯k\displaystyle=u^{k+1}+\beta(u^{k+1}-\bar{u}^{k})=(1+\beta)u^{k+1}-\beta\bar{u}^{k} (11)
=(1+β)uk+1β(uk+β(uku¯k1))\displaystyle=(1+\beta)u^{k+1}-\beta(u^{k}+\beta(u^{k}-\bar{u}^{k-1}))
=(1+β)uk+1β(1+β)uk+β2(1+β)uk1β3u¯k2\displaystyle=(1+\beta)u^{k+1}-\beta(1+\beta)u^{k}+\beta^{2}(1+\beta)u^{k-1}-\beta^{3}\bar{u}^{k-2}
=\displaystyle=\cdots\cdots
=(1+β)uk+1β(1+β)uk++(1)kβk(1+β)u1\displaystyle=(1+\beta)u^{k+1}-\beta(1+\beta)u^{k}+\cdots+(-1)^{k}\beta^{k}(1+\beta)u^{1}
+(1)k+1βk+1u0.\displaystyle~{}~{}~{}+(-1)^{k+1}\beta^{k+1}u^{0}.

Since {uk}\left\{u^{k}\right\} is bounded, there exists a constant N>0N>0 such that ukN\|u^{k}\|\leq N, k0\forall k\geq 0. If 0<β<5120<\beta<\frac{\sqrt{5}-1}{2} holds, then β(1+β)<1\beta(1+\beta)<1. It follows from (11) that

u¯k+1\displaystyle\|\bar{u}^{k+1}\| (1+β)uk+1+β(1+β)uk+\displaystyle\leq(1+\beta)\|u^{k+1}\|+\beta(1+\beta)\|u^{k}\|+\cdots
+βk(1+β)u1+βk+1u0\displaystyle~{}~{}~{}+\beta^{k}(1+\beta)\|u^{1}\|+\beta^{k+1}\|u^{0}\|
<(1+β)N+N+βN+β2N++βk1N+βkN\displaystyle<(1+\beta)N+N+\beta N+\beta^{2}N+\cdots+\beta^{k-1}N+\beta^{k}N
<(1+β)N+N(1+β+β2++βk1+βk)\displaystyle<(1+\beta)N+N(1+\beta+\beta^{2}+\cdots+\beta^{k-1}+\beta^{k})
=(1+β)N+Ni=1kβi\displaystyle=(1+\beta)N+N\sum_{i=1}^{k}\beta^{i}
<(1+β)N+Ni=1βi=(1+β)N+N1β\displaystyle<(1+\beta)N+N\sum_{i=1}^{\infty}\beta^{i}=(1+\beta)N+\frac{N}{1-\beta}
=2β21βN.\displaystyle=\frac{2-\beta^{2}}{1-\beta}N.

which consequently results in the boundedness of sequence {uk}\left\{u^{k}\right\}. By (6), we have (1+α)zk+1=z¯k+1+αz¯k(1+\alpha)z^{k+1}=\bar{z}^{k+1}+\alpha\bar{z}^{k}. Thus, {zk}\left\{z^{k}\right\} is bounded because {z¯k}\left\{\bar{z}^{k}\right\} is bounded. Therefore, for k0k\geq 0, we get {vk}\left\{v^{k}\right\} and {v¯k}\left\{\bar{v}^{k}\right\} are bounded. ∎

TABLE II: Comparison of results between different algorithms on the AAPM and Covid-19 datasets (MAE/RMSE). The best value of the traditional algorithm is marked with bold. The best value for deep learning is marked with red.
Data AAPM Dataset
Method Traditional algorithm Deep learning algorithm
Method FBP BM3D HQS-CG HQSp-CG PWLS-CSCGR IHQSp-CG PDNet FBPNet MetaInvNet IHQSp-Net* IHQSp-Net
Noise σ=0.3\sigma=0.3
180 13.18/16.79 10.61/14.66 4.73/6.74 4.51/6.41 7.98/453.28 4.48/6.37 5.69/8.07 4.74/6.67 4.17/5.79 4.29/5.99 4.07/ 5.68
120 17.45/23.56 11.97/17.45 5.51/8.01 5.10/7.48 8.06/438.85 5.03/7.37 6.33/9.22 5.34/7.78 4.52/ 6.37 4.83/6.89 4.54/6.42
90 19.63/25.21 12.24/16.96 6.25/9.26 5.62/8.50 9.12/503.83 5.50/8.29 6.57/9.90 5.76/8.48 4.78/6.82 5.28/7.69 4.70/ 6.74
60 24.54/31.38 14.36/19.48 7.61/11.62 6.57/10.43 11.07/619.98 6.36/10.11 7.56/11.94 6.48/10.12 5.37/ 7.79 6.08/9.20 5.45/7.90
Noise σ=0.3\sigma=0.3 and I0=5×105I_{0}=5\times 10^{5}
180 14.61/18.54 10.61/14.67 5.45/7.61 5.14/7.31 9.04/491.19 5.13/7.32 6.01/8.56 4.88/6.94 4.58/6.48 4.86/6.83 4.47/ 6.30
120 19.14/25.41 11.98/17.45 6.35/9.01 5.64/8.29 14.26/731.87 5.58/8.22 6.84/10.00 5.48/7.97 4.84/6.99 5.37/7.69 4.85/ 6.94
90 21.53/27.43 12.24/16.96 7.12/10.28 6.16/9.32 13.22/692.11 6.02/9.12 7.37/11.23 6.01/8.90 5.19/7.46 5.78/8.38 5.05/ 7.26
60 26.72/33.85 14.30/19.43 8.44/12.54 7.19/11.31 13.36/718.63 6.97/11.04 8.32/13.56 6.43/10.03 5.71/8.44 6.37/9.50 5.47/ 8.05
Data Covid-19 Dataset
Method FBP BM3D HQS-CG HQSp-CG PWLS-CSCGR IHQSp-CG PDNet FBPNet MetaInvNet IHQSp-Net* IHQSp-Net
Noise σ=0.3\sigma=0.3
180 12.88/16.26 10.49/13.64 4.03/5.62 3.77/5.27 6.86/289.66 3.73/5.19 5.50/7.50 4.14/5.68 3.54/4.78 3.65/4.97 3.42/ 4.65
120 16.75/22.56 11.69/15.79 4.75/6.84 4.33/6.32 7.16/296.62 4.21/6.07 6.06/8.50 4.75/6.76 3.77/ 5.19 4.19/5.85 3.85/5.32
90 18.60/24.11 12.01/15.91 5.51/8.23 4.86/7.44 7.99/342.59 4.69/7.06 6.15/8.99 5.15/7.46 4.04/5.69 4.58/6.57 3.94/ 5.52
60 23.30/30.45 14.40/19.70 6.98/11.05 5.93/9.81 9.81/432.80 5.61/9.19 7.08/11.03 6.13/9.80 4.54/ 6.62 5.41/8.06 4.57/6.66
Noise σ=0.3\sigma=0.3 and I0=5×105I_{0}=5\times 10^{5}
180 14.35/18.14 10.49/13.65 4.84/6.63 4.47/6.29 8.07/326.25 4.44/6.26 5.87/8.04 4.28/5.94 3.87/5.40 4.32/5.99 3.80/ 5.24
120 18.59/24.58 11.70/15.81 5.68/7.98 4.92/7.20 13.59/544.19 4.82/7.02 6.23/8.80 4.88/6.94 4.18/5.92 4.80/6.78 4.14/ 5.80
90 20.72/26.58 12.01/15.93 6.45/9.35 5.44/8.33 12.32/504.26 5.25/7.97 7.08/10.47 5.26/7.76 4.46/6.39 5.22/7.47 4.30/ 6.09
60 25.79/33.21 14.37/19.68 7.84/12.00 6.57/10.68 12.22/517.15 6.27/10.16 7.72/12.48 6.05/9.33 5.02/7.45 5.88/8.75 4.68/ 6.81
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
(a) Ground truth (b) FBP (c) BM3D (d) HQS-CG (e) HQSp-CG (f) PWLS-CSCGR
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
(g) IHQSp-CG (h) PDNet (i) FBPNet (j) MetaInvNet (k) IHQSp-Net* (l) IHQSp-Net
Figure 5: 60 sparse views CT image reconstruction results and magnified ROIs from AAPM. The sinogram is corrupted by σ=0.3\sigma=0.3.
Theorem VI.1.

Let {vk}\left\{v^{k}\right\} and {v¯k}\left\{\bar{v}^{k}\right\} be the sequences generated by our algorithm. Then any cluster point of {vk}\left\{v^{k}\right\} is the global minimum point of LL.

Proof.

By Lemmal VI.1, the sequences {vk}\left\{v^{k}\right\} and {v¯k}\left\{\bar{v}^{k}\right\} are bounded, which implies the existence of convergent subsequences, denoted as

vkjv,j,andv¯kjv¯,j.v^{k_{j}}\rightarrow v^{*},j\rightarrow\infty,\quad\mathrm{and}\quad\bar{v}^{k_{j}}\rightarrow\bar{v}^{*},j\rightarrow\infty.

And further, it can be proved from (6) and (8) that

{z¯k+1+αz¯k=(1+α)zk+1,u¯k+1+βu¯k=(1+β)uk+1.\left\{\begin{array}[]{ll}\bar{z}^{k+1}+\alpha\bar{z}^{k}=(1+\alpha)z^{k+1},\\ \bar{u}^{k+1}+\beta\bar{u}^{k}=(1+\beta)u^{k+1}.\\ \end{array}\right. (12)

Passing to the limit in (12) along the subsequences {vkj}\left\{v^{k_{j}}\right\} and {v¯kj}\left\{\bar{v}^{k_{j}}\right\} yields

{(1+α)z¯=(1+α)z,(1+β)u¯=(1+β)u,\left\{\begin{array}[]{ll}(1+\alpha)\bar{z}^{*}=(1+\alpha)z^{*},\\ (1+\beta)\bar{u}^{*}=(1+\beta)u^{*},\\ \end{array}\right. (13)

which means z¯=z\bar{z}^{*}=z^{*} and u¯=u\bar{u}^{*}=u^{*}. It follows from (5) and (7) that

{0λzk+1ppi=1mγi(Wiu¯kzik+1),0=AT(Auk+1f)+i=1mγiWiT(Wiuk+1z¯ik+1).\left\{\begin{array}[]{ll}0\in\lambda\partial\|z^{k+1}\|_{p}^{p}-\sum_{i=1}^{m}\gamma_{i}(W_{i}\bar{u}^{k}-z^{k+1}_{i}),\\ 0=A^{T}(Au^{k+1}-f)+\sum_{i=1}^{m}\gamma_{i}W_{i}^{T}(W_{i}u^{k+1}-\bar{z}^{k+1}_{i}).\\ \end{array}\right. (14)

Taking limits on both sides of (14) along the subsequences {vkj}\left\{v^{k_{j}}\right\} and {v¯kj}\left\{\bar{v}^{k_{j}}\right\}, when jj\rightarrow\infty and using closedness of subdifferential, we obtain

{0λzppi=1mγi(Wiuzi),0=AT(Auf)+i=1mγiWiT(Wiuzi).\left\{\begin{array}[]{ll}0\in\lambda\partial\|z^{*}\|_{p}^{p}-\sum_{i=1}^{m}\gamma_{i}(W_{i}{u}^{*}-z^{*}_{i}),\\ 0=A^{T}(Au^{*}-f)+\sum_{i=1}^{m}\gamma_{i}W_{i}^{T}(W_{i}u^{*}-{z}^{*}_{i}).\\ \end{array}\right. (15)

In particular, vv^{*} is a stationary point of LL. LL is a quasi-convex function, since zpp\|z\|_{p}^{p} is a quasi-convex function. Consequently, this property results in that v=(z,u)v^{*}=(z^{*},u^{*}) is the global minimum point of LL. ∎

VII Supplementary experimental data

VII-A Comparison of MAE and RMSE

Table II gives the comparison of MAE and RMSE for all the compared algorithms. The results in Table II are similar to those in Table I. IHQSpS_{p}-CG performs optimally in the traditional algorithm. IHQSp-Net also performs best among the deep learning algorithms in aggregate. This shows the robust results of the proposed IHQSp algorithm under different metrics.

VII-B Supplementary visualisation results

Fig. 5 illustrates a qualitative comparison of the 60-degree observation resulting from the reconstruction. It can be noticed that MetaInvNet is prone to pseudo-generation. HQS-CG, on the other hand, is not able to complete the reconstruction details. Overall, IHQSp-CG and IHQSp-Net are good at preserving details in both traditional and deep learning algorithms.