Self Supervised Low Dose Computed Tomography Image Denoising Using Invertible Network Exploiting Inter Slice Congruence
Abstract
The resurgence of deep neural networks has created an alternative pathway for low-dose computed tomography denoising by learning a nonlinear transformation function between low-dose CT (LDCT) and normal-dose CT (NDCT) image pairs. However, those paired LDCT and NDCT images are rarely available in the clinical environment, making deep neural network deployment infeasible. This study proposes a novel method for self-supervised low-dose CT denoising to alleviate the requirement of paired LDCT and NDCT images. Specifically, we have trained an invertible neural network to minimize the pixel-based mean square distance between a noisy slice and the average of its two immediate adjacent noisy slices. We have shown the aforementioned is similar to training a neural network to minimize the distance between clean NDCT and noisy LDCT image pairs. Again, during the reverse mapping of the invertible network, the output image is mapped to the original input image, similar to cycle consistency loss. Finally, the trained invertible network’s forward mapping is used for denoising LDCT images. Extensive experiments on two publicly available datasets showed that our method performs favourably against other existing unsupervised methods.

1 Introduction
Low dose computed tomography (LDCT) imaging offers a nonvulnerable solution to the frequent requirement of computed tomography (CT) imaging by reducing the patient’s radiation exposure at the time of imaging [2], [22]. However, the lowering of radiation dose leads to degradation of the diagnostic quality of the low dose CT image, by creating several artifacts and noise. In the last few decades, researchers have tried to restore the diagnostic quality of LDCT images by various image post-processing methods. Among those methods, the deep neural network have been proven to be most efficient due to its commendable expressive power [3]. Despite the superior image quality achieved by neural networks as compared to handcrafted prior functions, almost all the neural network-based methods require both noisy and clean images during the training. However, paired clean and noisy images are rarely available in practical scenarios. So, training deep networks without clean images is of crucial importance. Recently, there has been a surge in the studies exploring the possibilities of training deep networks without clean images for the purpose of denoising natural images corrupted with synthetic noise [16], [1], [23]. However, we found these generalized solutions do not perform adequately in complicated signal-dependent CT noise or are ill-suited because of inconvenient training procedures. Recently Wu et al. [25] proposed a self supervised method for LDCT denoising based on the Noise2Noise [16] paradigm. However, in order to generate two independent realization of the same noisy image, their methods requires to divide the whole projection data into odd and even projection set. This heuristic approach lowers spatial resolution and signal strength of the input image, and also restricts the deployment of the technique due to interference with the workflow of existing CT scanners.
In this study, we have proposed a novel method for self supervised low dose CT denoising. Firstly, we have proposed to use the inter slice congruence for generating two noisy observation of the same image. We used the axial mean of the two opposite neighbouring slices to approximate the second noisy observation of the contemporary slice. However, the above mentioned method created a residual offset between two noisy observations. As a remedy we have leveraged the recently proposed concept of invertible neural network, which act both as a regularizer and a restoration network. Specifically, during the forward mapping we used the network to map a noisy axial slice to mean of its two opposite neighbouring slices, and during the reverse mapping the output is again mapped back to the original noisy image. The information lossless property of the invertible network work as a regularizer to overcome the residual offset which occurs due to axial approximation.
We evaluated our proposed method on two datasets; namely, the 2016 NIH-AAPM-Mayo Clinic Low Dose CT Grand Challenge, and the ELCAP Public Lung Image Database. Extensive experiments on both these data sets demonstrated that our method is notably superior to other state-of-the-art unsupervised or self supervised low-dose CT denoising methods.
2 Comparison with other methods:
In recent times a lot of attention has been given to train neural network without clean images pairs. Among these methods, void networks or noise2self [13], [1], [15] are the leaders for their performance in synthetic noise denoising, but we have shown these networks do not perform adequately in complicated CT noise. On the other hand, deep image prior [23] is another self-supervised method to denoise image. However, the iterative deep image prior doesn’t have any defined stopping criteria, which makes it unsuitable for medical applications; as with more iteration the network produces the original noisy image. The CycleGAN framework [29] serves as the foundation for another well-known school of unsupervised denoising methods [17], [18], [9]. Unpaired training data were utilized in these methods to train the denoising network. However, it is difficult and laborious to converge the networks trained with the CycleGAN framework, which makes these approaches inconvenient for widespread use. The chosen hyperparameter has an impact on denoising performance as well. Recently, D Wu et al. [25] proposed a consensus loss function for LDCT denoising; however, their methods require dividing the projection data into two sets before back-projection of sinogram data to reconstruct the input image. This restricts the deployment of the technique due to interference with the workflow of existing CT scanners. Whereas, our method is a simple plug-and-play method and does not interfere with workflow of the standard CT scanners.
3 Method
3.1 Invertible Neural Network
Invertible neural networks(INN) are originally designed for unsupervised learning of probabilistic models [5]. These networks can transform a distribution to another distribution through a bijective function without losing information. Therefore, they are ideally suited to be used as normalizing flow [12]. In recent work they have been used in various image to image translation task [28], [24], image denoising [14], [20], image re-scaling [26], image super resolution [19], etc.
INNs consist of layers that guarantee an invertible relationship between their input and output. Every invertible block, must satisfy the following condition, if , then , where is the invertible block and and are the input and output to the block. In our study, we leveraged the invertible blocks formulation proposed in [6]. For the block the input is split into two groups and using channel wise splitting of same size. Then, they undergo an additive affine transformation with an identity branch augmentation as follows:
(1) |
(2) |
Here, implies element wise multiplication. During the reverse mapping, given the output and , the input and can computed as
(3) |
(4) |
Here, , , can be any function(e.g., convolutional layer) without any restriction on invertibility. The above formulation enable to directly determine inverse function without any complicated computation or information loss.
3.2 Inter Slice Congruence
Let’s consider , , are three adjacent noisy slices, and , , are the corresponding clean CT slices. In general, , where is the random
variable of observed noise. We do have any pre-assumption on the characteristic of the observed noise. The “” sign signifies that there are observed noise at every pixel added with original signal. Our only assumption is , , and , i.e., the observed noise in different axial slices are statistically independent from each other. Which is always true for CT images.
Ideally,
.
Where is the offset which compensates for the small axial distance between two slices. Substituting the value of the we can get, , or .
Similarly from , we can get .
Adding the above two expressions, we get
(5) |
or,
(6) |
Here, . The value of is very small as compared to the , so we can ignore this term (we will discuss more about it in section 2.3). A deep neural network , parameterized by , can be trained to minimize the following distance function:
(7) |
or,
(8) |
or,
(9) |
According to Lehtinen et al. [16], training a neural network to minimize the distance between two independent noisy realizations of the same clean signal is equivalent to training a neural network by minimizing the distance between the original clean signal and noisy signal. So,
(10) |
Layer Name | Details |
---|---|
Single Covolutional Layer with 1 input channel and 64 output channel. Kernel size . | |
Single Covolutional Layer with 1 input channel and 64 output channel. Kernel size . | |
Single Covolutional Layer with 64 input channel and 1 output channel. Kernel size . | |
Single Covolutional Layer with 64 input channel and 1 output channel. Kernel size . | |
Dense Block [8] with grow rate of 32, and four layer. Kernel of size , 64 input and output channel. |

3.3 Design Concept
We have given a graphical representation of our proposed method in Figure 1. As shown in the Figure, the noisy input image is mapped to the average of its two neighboring slices during the forward mapping. As discussed in the above section, it is similar to training a neural network to map a noisy image to the noise-free image counterpart. Although we said the offset is negligible, but in some cases, it may become considerable, especially at the beginning and the end of the organ (we have shown an example in the section 4.1). To regularize the network in this end organ case, we again mapped the output image back to the noisy input image during the reverse mapping of the invertible network, similar to cycle consistency loss. Furthermore, the architectures of the invertible network by design satisfy cycle-consistency before training even begins, which again regulates the network from the interslice smoothing at the end of organ cases. In our design, we employed two encoders (,), two decoders (,), and a invertible core . The encoder takes the input noisy slice as the input and produces a higher dimension overcomplete 3D feature map , i.e., . Here, and are the width and the height of the input noisy image patch, and is the number of channels in the feature map. This feature map is used as the input to the invertible core during the forward map. The invertible core produces an output . The decoder projects the higher dimensional feature map to image space. Thus, the output of the decoder is the predicted clean image , i.e., . During the reverse mapping, the encoder first takes the predicted clean image as the input and produces a higher dimension overcomplete 3D feature map , i.e., . Then, this feature map is transformed into a higher dimensional feature map using the reverse invertible core . Finally the decoder project into predicted noisy image . So the mapping for the forward and reverse processes are given by
The invertible core and its inverse shares parameters with each other. Essentially training the will automatically train and vise versa. Via training the invertible network for reverse mapping we are regulating the information loss which may happen in the forward mapping in the end organ cases, as for the reverse mapping we have exact pixel by pixel correspondence between the target noisy image and predicted noisy image.
A visual illustration of the proposed network is given in Figure 2. We used pixel shuffle and unshuffle layers to downsample and upsample the feature maps, as these two layers satisfy the invertibility property. We used total invertibles blocks in total. A detailed layer-wise description of the network is given in Table 1.
3.4 Learning
We used mean square error as the reconstruction loss in our study. In the forward mapping the loss is calculated between the predicted clean image and average of two neighbouring noisy slice, i.e.
(11) |
or,
(12) |
During reverse mapping the loss is calculated between the predicted noisy image and original noisy image , i.e.
(13) |
or,
(14) |
or,
(15) |
Here is the total number of samples in every batch. The total loss is back-propagated for each batch to train the network. Note that the loss never minimizes completely during the training, as here the network is asked to solve an impossible task, i.e., transform a noisy version of an image into another noisy version. After training, the network produces images with the smallest average deviation from the noisy target images, and identical to the general class of deviation-minimizing estimator [10], [16]; this estimated image turns out to be the original clean image.














4 Experimental Settings
We evaluated the proposed method on one synthesized dataset and one clinical datasetst referred as D1, and D2, described below: 1. D1:2016 NIH-AAPM-Mayo Clinic Low Dose CT Grand Challenge Dataset 2. D2:ELCAP Public Lung Image Database. The first dataset contains simulated quarter dose CT scan of 10 patients. For each patient, there are approximately 250 slices per scan. The second data set includes real low-dose CT scans from 50 patients. We used Adam optimizer with a batch size of 16. The learning rate was initially set to and was set to decrease by a factor of 2 after every 6000 iterations. For training, we have taken ten randomly cropped patches from each slice, each patch of size . The benchmark data set D1 contains paired NDCT and LDCT images, so we chose this dataset for the objective evaluation of our proposed method, and D2 is used for subjective evaluation only. For dataset D1, we used CT slices from seven patients as the training set and used the remaining three patients’ data as the test set. For dataset D2, we used images from 30 patients as our training data and followed the same training procedure as mentioned above.
5 Result and Discussion
5.1 Ablation Study
This section systemically investigates the efficacy of every module proposed in this study. We considered three different networks; first, baseline model(M1), where the inverting block is replaced with dense block, and trained using minimizing mean square distance between two noisy observations(i.e. ). Next, two independent baseline model(one for forward mapping, one for reverse mapping) is jointly trained using a linear combination , and , similar to cycle consistent network paradigm. The forward mapping network is used for testing. We refer this model as M2. In both M1, and M2, we increased the depth of the network to make the representation power of these networks comparable with inverting network. Finally, the proposed method, referred as M3. Table 2 depicts the objective evaluation of the three networks using the D1 dataset. Both M2 and M3 use reverse mapping to regularize the network; the influence of the same in the denoising performance is evident from Table 2. Adding cycle consistency loss has improved the performance of the same baseline model significantly. The inverting network performed considerably better than the network M2. It improves PSNR by 0.23dB. As discussed in the above section, in case cycle consistency loss, an additional network is trained, but that does not always guarantee invertibility, whereas inverting network architecture inherently possesses reversibility, which acts as a strong regularizer. In Figure 3 we have shown denoising performance of different networks visually. To demonstrate the requirement of regularization, we first extract the boundary line of the various organ from the NDCT image and superimpose the boundary line on the output of different networks. As shown in Figure 3, many pixels around the boundary line of M1 network output are missing. Using the reverse mapping, the issue of the end-organ missing pixel is successfully overcome in M2 and M3. The granular pattern is also less present in the M3 than M2. The zoomed version of a ROI taken from the images of Figure 3 is given in Figure 4 for better perception. In Figure 5 we give an example of the performance of the invertible network in reverse mapping. Here, the predicted LDCT image is produced by using the predicted clean image of the forward mapping as the input for reverse mapping. As shown, the predicted noisy pattern is similar to the original noise pattern. The same streaking artifacts are present in both the noise pattern; also, the noise variance is different in the various spatial region depending on the signal intensity of the original CT image. It validates that the loss of information in reverse mapping is minimal. Due to the invertible network’s structural advantage, the network also preserves every information present in the input image in the forward mapping.
5.2 Performance on the simulated data
We compare performance of our method against four state of the art unsupervised method; namely, DIP [23], ConsensusNet [25], CycleGAN [17], BM3D [4]. Performance metrics used are, peak signal to noise ratio (PSNR), structural similarity index measure (SSIM), fréchet inception distance (FID) [7], perceptual loss (PL) [11], texture matching loss (TML) [21], and visual information fidelity (VIF) [27].
Method | PSNR | SSIM |
---|---|---|
M1 | 30.80 | 0.891 |
M2 | 31.03 | 0.899 |
M3 | 31.26 | 0.908 |
Method | PSNR | SSIM | FID | TML | PL | VIF |
---|---|---|---|---|---|---|
DIP | 28.08 | 0.848 | 58.96 | 0.402 | 0.00189 | 0.652 |
ConsensusNet | 29.02 | 0.881 | 39.47 | 0.079 | 0.00171 | 0.711 |
BM3D | 30.02 | 0.885 | 49.23 | 0.375 | 0.00152 | 0.732 |
CycleGAN | 30.86 | 0.886 | 30.85 | 0.038 | 0.00054 | 0.955 |
Proposed | 31.26 | 0.893 | 30.23 | 0.012 | 0.00068 | 0.989 |
As shown in Table 3, DIP has exhibited the worst performance among all the methods. On the other hand, the ConsensusNet yielded a better FID and TML than BM3D but a lower PSNR and SSIM than other methods. The ConsensusNet divided the original projection data of the low dose CT image into two subgroups and back-projected to create the noisy input signal. Consequently, the noisy input image is much noisier than the original LDCT image. Also, the structural loss occurred during the generation of noisy images, so as a result, the PSNR and SSIM of this method are lower than other methods. Texture matching loss (TML) is used to measure the texture difference between the reconstructed and original images. The lower value of TML indicates that the generated texture is similar to the original. In comparison, FID estimates the distance between the distribution of the generated image and real images. A lower value of FID signifies the generated images are more similar to the original image. The current deep learning era demands a denoised image with a low value of these metrics. These denoised images may be used as input for other image classification tasks or segmentation networks. In this regard, the ConsensusNet is superior to the BM3D because it uses the deep neural network’s expression power. CycleGAN is another powerful unsupervised method for image-to-image translation; it achieved better performance than the other methods. However, CycleGAN has a lot of bottlenecks, e.g., longer training time, computation power, hyper-parameter tuning, etc. All these bottlenecks make CycleGAN ill-suited for practical deployment. Meanwhile, our proposed method has achieved the highest PSNR, SSIM, FID, and VIF among all the other methods. Next, we compare the result of denoising visually in Figure 6. It can be observed that the proposed method performs significantly better than the other unsupervised methods. BM3D output produced a blurry denoised image and contained many splotchy artifacts. The same blurriness can be observed in the output of ConsensusNet, and DIP, although noise suppression is adequate, and splotchy artifact is absent. In the output of CycleGAN, we can observe the presence of residual noise, especially in the high noise regions. Next, we identified one low attenuated lesion in the sample image and marked the lesion with a red colour bounding box. The zoomed view of the region inside the bounding box is given in Figure 7. In our method’s output image, the lesion’s visibility is enhanced significantly than in other methods. Despite being an unsupervised method, the visibility of the lesion is comparable with the original NDCT image. Also, from the zoomed view, we can perceive that our method has suppressed the granular pattern without losing the original image’s texture.


























5.3 Performance on the real low dose CT data
Here, we analyse the performance of the proposed method on clinical data. Since there are no ground truths available, only qualitative comparisons are performed. An example of the denoising performance of the proposed method is given in Figure 8. Here we can see denoising performance of BM3D is the worst compared to other methods. On the other hand, the output of the deep image prior is very blurry, different organ boundary is distorted, and some splotchy artifacts appear in the image. The performance of DIP depends mainly on the stopping iteration; with more iterations, it will again produce the original noisy image. The clean image produced by it is always blurry and without any texture. Consistent with our previous example, we can see CycleGAN has a lot of residual noise left in the example. Next, we identified one hypodense lesion in the image and marked the region with the red colour bounding box. The zoomed version of the region containing the lesion is shown in Figure 9. The effect of denoising is perceptible in this zoomed view. The visibility of the lesion is very inconspicuous in the original LDCT image. The noise variance is very high in this region; consequently, BM3D and CycleGAN failed to remove noise from this region. On the other hand, DIP has removed the noise but also destroyed the image by removing all the texture information. ConsensusNet also produced a blurry version of the lesion in the output. At the same time, our method produced a denoised image with the lowest granular pattern and improved the lesion’s visibility. The main objective of image denoising is to restore the visibility of these types of lesions and anomalies by concealing the noise. In this regard, our method has reached the goal as the perceptibility of different anomalies has been improved vastly without losing structural or textural information.
6 Conclusion
This study proposed a novel method for self-supervised LDCT denoising. The proposed method is built upon the well-celebrated Noise2Noise paradigm. We have proposed an innovative way to produce the second noisy image for training denoiser via taking the average of the two neighbouring slices of each slice. The learnt denoiser is then improved by leveraging the recent concept of invertible learnable blocks and adding a cycle consistency loss in the training stage. The proposed method is an elegant way to use inter-slice correlation for LDCT denoising. Moreover, the proposed method is straightforward to deploy and does not interfere with the workflow of the existing CT scanner. Comprehensive evaluation on both synthetic low dose and real low dose CT data validates our method’s superiority over other self-supervised methods.
References
- [1] Joshua Batson and Loic Royer. Noise2self: Blind denoising by self-supervision. In International Conference on Machine Learning, pages 524–533. PMLR, 2019.
- [2] David J Brenner. Radiation risks potentially associated with low-dose ct screening of adult smokers for lung cancer. Radiology, 231(2):440–445, 2004.
- [3] Hu Chen, Yi Zhang, Mannudeep K Kalra, Feng Lin, Yang Chen, Peixi Liao, Jiliu Zhou, and Ge Wang. Low-dose ct with a residual encoder-decoder convolutional neural network. IEEE transactions on medical imaging, 36(12):2524–2535, 2017.
- [4] Kostadin Dabov, Alessandro Foi, Vladimir Katkovnik, and Karen Egiazarian. Image denoising with block-matching and 3d filtering. In Image Processing: Algorithms and Systems, Neural Networks, and Machine Learning, volume 6064, page 606414. International Society for Optics and Photonics, 2006.
- [5] Laurent Dinh, David Krueger, and Yoshua Bengio. Nice: Non-linear independent components estimation. arXiv preprint arXiv:1410.8516, 2014.
- [6] Laurent Dinh, Jascha Sohl-Dickstein, and Samy Bengio. Density estimation using real nvp. arXiv preprint arXiv:1605.08803, 2016.
- [7] Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017.
- [8] Gao Huang, Zhuang Liu, Laurens Van Der Maaten, and Kilian Q Weinberger. Densely connected convolutional networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4700–4708, 2017.
- [9] Zhiyuan Huang, Zixiang Chen, Qiyang Zhang, Guotao Quan, Min Ji, Chengjin Zhang, Yongfeng Yang, Xin Liu, Dong Liang, Hairong Zheng, et al. Cagan: A cycle-consistent generative adversarial network with attention for low-dose ct imaging. IEEE Transactions on Computational Imaging, 6:1203–1218, 2020.
- [10] Peter J Huber. Robust estimation of a location parameter. In Breakthroughs in statistics, pages 492–518. Springer, 1992.
- [11] Justin Johnson, Alexandre Alahi, and Li Fei-Fei. Perceptual losses for real-time style transfer and super-resolution. In European conference on computer vision, pages 694–711. Springer, 2016.
- [12] Diederik P Kingma and Prafulla Dhariwal. Glow: Generative flow with invertible 1x1 convolutions. arXiv preprint arXiv:1807.03039, 2018.
- [13] Alexander Krull, Tim-Oliver Buchholz, and Florian Jug. Noise2void-learning denoising from single noisy images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2129–2137, 2019.
- [14] Taesung Kwon and Jong Chul Ye. Cycle-free cyclegan using invertible generator for unsupervised low-dose ct denoising. arXiv preprint arXiv:2104.08538, 2021.
- [15] Samuli Laine, Tero Karras, Jaakko Lehtinen, and Timo Aila. High-quality self-supervised deep image denoising. arXiv preprint arXiv:1901.10277, 2019.
- [16] Jaakko Lehtinen, Jacob Munkberg, Jon Hasselgren, Samuli Laine, Tero Karras, Miika Aittala, and Timo Aila. Noise2noise: Learning image restoration without clean data. arXiv preprint arXiv:1803.04189, 2018.
- [17] Zeheng Li, Junzhou Huang, Lifeng Yu, Yujie Chi, and Mingwu Jin. Low-dose ct image denoising using cycle-consistent adversarial networks. In 2019 IEEE Nuclear Science Symposium and Medical Imaging Conference (NSS/MIC), pages 1–3, 2019.
- [18] Zeheng Li, Shiwei Zhou, Junzhou Huang, Lifeng Yu, and Mingwu Jin. Investigation of low-dose ct image denoising using unpaired deep learning methods. IEEE Transactions on Radiation and Plasma Medical Sciences, 5(2):224–234, 2020.
- [19] Jingyun Liang, Andreas Lugmayr, Kai Zhang, Martin Danelljan, Luc Van Gool, and Radu Timofte. Hierarchical conditional flow: A unified framework for image super-resolution and image rescaling. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 4076–4085, 2021.
- [20] Yang Liu, Zhenyue Qin, Saeed Anwar, Pan Ji, Dongwoo Kim, Sabrina Caldwell, and Tom Gedeon. Invertible denoising network: A light solution for real noise removal. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 13365–13374, 2021.
- [21] Mehdi SM Sajjadi, Bernhard Scholkopf, and Michael Hirsch. Enhancenet: Single image super-resolution through automated texture synthesis. In Proceedings of the IEEE International Conference on Computer Vision, pages 4491–4500, 2017.
- [22] Alvin W Su, Travis J Hillen, Eric P Eutsler, Asheesh Bedi, James R Ross, Christopher M Larson, John C Clohisy, and Jeffrey J Nepple. Low-dose computed tomography reduces radiation exposure by 90% compared with traditional computed tomography among patients undergoing hip-preservation surgery. Arthroscopy: The Journal of Arthroscopic & Related Surgery, 35(5):1385–1392, 2019.
- [23] Dmitry Ulyanov, Andrea Vedaldi, and Victor Lempitsky. Deep image prior. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 9446–9454, 2018.
- [24] Tycho FA van der Ouderaa and Daniel E Worrall. Reversible gans for memory-efficient image-to-image translation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4720–4728, 2019.
- [25] Dufan Wu, Kuang Gong, Kyungsang Kim, Xiang Li, and Quanzheng Li. Consensus neural network for medical imaging denoising with only noisy training samples. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 741–749. Springer, 2019.
- [26] Mingqing Xiao, Shuxin Zheng, Chang Liu, Yaolong Wang, Di He, Guolin Ke, Jiang Bian, Zhouchen Lin, and Tie-Yan Liu. Invertible image rescaling. In European Conference on Computer Vision, pages 126–144. Springer, 2020.
- [27] Guorong Xuan, Chengyun Yang, Yizhan Zhen, Yun Q Shi, and Zhicheng Ni. Reversible data hiding based on wavelet spread spectrum. In IEEE 6th Workshop on Multimedia Signal Processing, 2004., pages 211–214. IEEE, 2004.
- [28] Rui Zhao, Tianshan Liu, Jun Xiao, Daniel PK Lun, and Kin-Man Lam. Invertible image decolorization. IEEE Transactions on Image Processing, 30:6081–6095, 2021.
- [29] Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A Efros. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE international conference on computer vision, pages 2223–2232, 2017.