Exploring the Power of Generative Deep Learning for Image-to-Image Translation and MRI Reconstruction: A Cross-Domain Review
Abstract
Deep learning has become a prominent computational modeling tool in the areas of computer vision and image processing in recent years. This research comprehensively analyzes the different deep-learning methods used for image-to-image translation and reconstruction in the natural and medical imaging domains. We examine the famous deep learning frameworks, such as convolutional neural networks and generative adversarial networks, and their variants, delving into the fundamental principles and difficulties of each. In the field of natural computer vision, we investigate the development and extension of various deep-learning generative models. In comparison, we investigate the possible applications of deep learning to generative medical imaging problems, including medical image translation, MRI reconstruction, and multi-contrast MRI synthesis. This thorough review provides scholars and practitioners in the areas of generative computer vision and medical imaging with useful insights for summarizing past works and getting insight into future research paths.
1 Introduction
The integration of deep learning techniques has enabled the application of convolutional neural networks (CNNs) for efficient computer vision. CNNs have proven to be effective in various computer vision tasks, including image classification [40], semantic segmentation [52], object detection [70] and medical image analysis [71], [45]. Another promising deep learning technique for generating new data through unsupervised and supervised learning of complex distributions is generative adversarial networks (GANs) [23]. GANs consist of two neural networks, a generator, and a discriminator, that work in an adversarial manner. One of the key research directions for GANs is natural image-to-image (I2I) translation and synthesis, which is the process of converting one image into another while preserving or changing its complex and high-level semantic content. The purpose of I2I translation is to enable machines to comprehend and generate realistic images that are similar to the input images but have distinct visual qualities. I2I translation is an important AI-related direction that can reduce a large amount of human labor in certain works in daily life and the medical domain. As we already know, deep CNN models can extract high-level image features; therefore, GAN-based models combined with CNN in the generator can recover learned features from unprocessed data. I2I translation includes multiple directions in the domain of natural imaging, including style transfer [53],[38],[100], image and video colorization [106],[108],[82], image Super-Resolution [112],[101], etc. GANs have been increasingly applied to medical image translation tasks [88] and have produced results comparable to conventional I2I translations. They are gradually replacing traditional non-generative deep learning approaches in various medical image analysis tasks involving different modalities, such as magnetic resonance imaging (MRI), computerized tomography (CT), and X-ray [27].
MRI is a cutting-edge medical imaging technique that has revolutionized medicine, radiology, and psychiatry [27]. With its ability to provide quantifiable and reproducible information about the structural, anatomical, and functional properties of tissue, MRI has surpassed CT scans and X-rays in terms of image clarity and patient safety, as it produces clearer images of organs and soft tissues without exposing the patient to radiation. However, the high complexity of the MRI imaging process can result in prolonged acquisition times due to the long process of generating high-resolution MR images from fully-sampled k-space. In contrast to conventional medical I2I translation, MRI reconstruction seeks to overcome this issue by creating de-aliased MR images using under-sampled k-space rather than fully-sampled k-space. This approach preserves the original signal for greater clinical usage while reducing the acquisition duration of an MR image. Compressed sensing (CS), a mathematical theory used for signal acquisition and processing, has been commonly used for this purpose [54]. Although reconstructing full-sampled images from under-sampled k-space is an ill-posed linear inverse task, incorporating assumptions or prior information, such as the sparsity of the frequency domain (k-space) of the MR image, can help regularize the solution by formulating a convex optimization problem. Methods such as SENSE [68] and GRAPPA [24], based on CS-MRI, have achieved remarkable success in the past decades.
As deep learning techniques gained popularity in the computer vision area, researchers sought to apply these methods to MRI reconstruction instead of relying on traditional CS-MRI methods. CNNs were able to learn a mapping from under-sampled k-space to the image domain, which enabled them to reconstruct the original MR image. Despite the automatic learning process in CNNs, the significant manual effort was still required to design effective losses. For instance, in some generative tasks, CNNs faced challenges in achieving good performance due to the difficulty in defining a suitable loss function. In contrast, GANs can be used to specify only a high-level goal, such as ”make the output indistinguishable from reality,” and then automatically learn an appropriate loss function to meet that goal [33]. As with other areas, researchers in the field of MRI reconstruction began utilizing GANs instead of CNNs, leading to substantial advancements. Nevertheless, medical imaging still presents numerous difficulties compared to natural imaging-based tasks. One major challenge is the scarcity of large datasets like ImageNet [73] for medical imaging, which is largely due to data privacy restrictions and protection policies. From a technical standpoint, medical imaging has physiological and anatomical significance, making the learning process more complex and requiring additional interpretation. This adds an extra layer of complexity to the field of medical imaging, compared to natural imaging.
The purpose of our survey is to provide a comparative overview of GAN-based models in I2I translation and MRI reconstruction. In Section II, we present a background on GANs, I2I Translation, and CS-MRI reconstruction. Section III reviews the literature on I2I Translation using deep learning models and provides in-depth information about Pix2Pix. Section IV focuses on the most well-known and benchmarked GAN-based deep learning models for MRI reconstruction. Section V explores more complex medical imaging reconstruction problems, such as multi-contrast MRI reconstruction. In Section VI, we outline commonly used evaluation metrics in I2I translation and MRI reconstruction and provide important considerations for conducting experiments in these fields. Finally, in Section VII, we discuss the challenges in I2I translation and MRI reconstruction and suggest potential future directions for research.
2 Background
2.1 Generative Adversarial Network
The Generative Adversarial Network (GAN) [23] has proven to be effective in several applications, such as I2I translation, image super-resolution, image reconstruction, and data augmentation. GANs are inspired by two-player zero-sum games in game theory and consist of two components: the generator and the discriminator . The generator is designed to learn the mapping from a noise distribution to the data space , while the discriminator operates as a binary classifier that distinguishes between the generated fake sample from the distribution and the real sample from the distribution . The loss function of the GAN model can be expressed as:
(1) | |||
Where is the expected value over all sampled real data instances, is the expected value over all generated fake instances . Training a GAN model involves simultaneously training the generator and discriminator . Given a batch of noise samples from the noise distribution and a batch of data samples from the data space , the discriminator is updated using its stochastic gradient ascent:
(2) |
After updating the discriminator , we need to update the generator by:
(3) |
Where and represents the trainable parameters from discriminator and generator .
2.2 Conditional Generative Adversarial Network
The Conditional Generative Adversarial Network (CGAN) [59] is a GAN variation incorporating a conditional element. In a CGAN, the generator and discriminator are conditioned on additional input data, such as an image or a label. This allows the CGAN to generate samples that are conditioned on specific characteristics, such as specific object classes in an image or specific attributes in a dataset. The objective function of a CGAN can be written as follows:
(4) | |||
The condition could be a label or an image, is the random noise, and represents the ground-truth image. The generator takes the condition and the random noise as input, while the discriminator receives the data pairs (, , )) and (, ) and then determines which is real and which is fake. Figure 1 illustrates the architecture of both GANs and CGANs.

2.3 CS-MRI reconstruction

CS-MRI (Compressed Sensing MRI) reconstruction is a method for acquiring MRI images with reduced data acquisition time, which can be particularly useful in applications where prolonged scanning times are not feasible, such as with patients who have difficulty remaining still for extended periods of time [34]. CS-MRI operates by sparsely sampling the k-space data, the frequency domain representation of the MRI image, i.e., acquiring a small number of samples. This produces an under-sampled k-space, which can then be reconstructed using advanced mathematical algorithms, such as sparse optimization algorithms or deep learning-based methods, to produce an image. Certain image features can be reconstructed from a sparse set of k-space samples, reducing the amount of data needed to obtain an acceptable image quality. This can lead to faster scan times, enhanced patient comfort, and cost savings.
Given a MR image , we do a fully-sampled measurement on k-space, using the Fourier transform :
(5) |
Where represents noise. Using compressed sensing, we can reconstruct from undersampled k-space measurement due to the sparsity characteristic of the MR image in its frequency domain. So that:
(6) |
if we use a binary sampling mask is a method for selecting a subset of k-space lines in order to accelerate the measurement, then:
(7) |
The sampling mask and the Fourier transform can be combined as an encoding matrix , in order to reconstruct , we need to minimize the noise, so that:
(8) |
Instead of using compressed sensing, non-generative DL-based approaches, such as CNNs, can be formulated as an optimization problem:
(9) |
Where represents the zero-filled reconstruction, is the output image from the trained model, is the optimized parameters in the trained model. Figure 2 shows the description of CS-based MRI reconstruction.
However, CS-MRI reconstruction has two major limitations: it requires a significant amount of computational time, and it is heavily dependent on the choice of regularization parameters and transforms basis, which must be adjusted carefully [56].

3 Image-to-Image Translation with Deep Learning
I2I translation is a subfield of computer vision and deep learning that seeks to convert images from one domain to another while maintaining their content and structure. Numerous tasks, including style transfer, semantic segmentation, object detection, image synthesis, and data augmentation, have found extensive applications for this technology. The continuous development of I2I translation promises to uncover additional potential applications in a variety of industries, such as the entertainment industry, healthcare, and autonomous vehicles. This section begins with a review of the literature on I2I translation using deep learning techniques. It also introduces two well-known GANs, Pix2Pix and CycleGAN, which are primarily concerned with image style transfer.
3.1 Related Works:
Previous works have shown the effectiveness of CGAN for I2I translation tasks. One notable example is the Pix2Pix model [33], which utilizes a CGAN to produce remarkable results in various applications, such as transforming satellite views into street maps and generating shoes from border sketches. The generator in Pix2Pix has a U-Net-like architecture that allows it to perform style transfer based on paired training images while preserving the pixel-wise characteristics of the source image. This results in accurate I2I translations combining high-level semantic and detailed pixel information. Zhu et al. [111] introduced CycleGAN, which used unpaired images as training samples combined with a cycle-consistency loss, producing more excellent results on image art style transfer. Wang et al. [91] proposed a new CGAN, HDPix2Pix, for high-resolution image synthesis, which can generate photo-realistic images without the need for hand-crafted losses or pre-trained networks. Almahairi et al. [2] then developed Augmental CycleGAN, which enables many-to-many mapping using unpaired data. Huang et al. [31] proposed MUNIT, a framework based on unsupervised I2I translation. MUNIT has outperformed CycleGAN in various I2I tasks. However, these related works to Pix2Pix have limitations in scalability and robustness when dealing with multiple domains. To address this, Choi et al. [13] proposed StarGAN, which can perform I2I translations for multiple domains using a single model. A ’domain’ is defined as images sharing the same attribute value, such as images of women or people with black hair. StarGAN v2 [14] is an updated version of the original StarGAN model [13]. StarGAN v2 uses the Adaptive Layer-Instance Normalization (AdaLIN) generator architecture, allowing the model to adjust the normalization parameters for each instance, improving performance for diverse domains. Additionally, StarGAN v2 has the capability to perform I2I translations in an unsupervised manner using unpaired training data. More recently, Cao et al. [9] introduced the Remix framework, which overcomes the over-fitting problem encountered with fewer training examples/paired images. The ReMix method can easily be incorporated into existing GAN models with minimal modifications. Table I shows some important GAN architectures for I2I translation.
Reference | Year | Dataset | Domain | highlight |
---|---|---|---|---|
Pix2Pix [33] | 2017 | Paired data | Two | First well-known GAN for I2I translation |
CycleGAN [111] | 2017 | unpaired data | Two | unpaired domain transfer |
Dual-GAN [100] | 2017 | unpaired data | Two | unpaired domain transfer |
HDpix2pix [91] | 2018 | paired data | Two | high-resolution improved Pix2Pix |
Attention GAN [11] | 2018 | unpaired data | Multiple | Attention-based |
StarGAN [13] | 2018 | unpaired data | Multiple | Multiple domains translation |
TravelGAN [3] | 2019 | unpaired data | Two | Vector learning instead of cycle consistency |
SelectionGAN [85] | 2019 | paired data | Two | Cross-view image translation |
StarGAN v2 [14] | 2020 | unpaired data | Multiple | Multiple domains translation |
SPA-GAN [21] | 2020 | unpaired data | Multiple | Spatial attention |
3.2 Pix2Pix
Pix2Pix is a pioneering and groundbreaking I2I translation infrastructure. It is a large-scale, systematic I2I project that implemented supervised paired data training for two-domain image translation using an image-based conditional GAN. Pix2Pix employs U-net as the generator backbone to preserve low-level and high-level structural and textural information and PatchGAN as the discriminator capable of handling large images based on the fixed-size patch discriminator. Pix2Pix has been trained on a number of paired image datasets, including the Cityscapes dataset for semantic label vs. photo, the CMP Facades dataset for architectural label vs. photo, map vs. aerial photo from Google Map, and sketch and edge vs. photo. These models performed well during validation and testing and also provide an online resource for anyone to conduct their own trials.
General definitions: Pix2Pix uses an image-CGAN architecture that consists of two networks: a generator and a discriminator . The aim of Pix2Pix is to translate an input image to a target image by learning a mapping and distribution based on a condition image. The generator network takes a conditional input image and a random noise as input. It generates a corresponding output image , while the discriminator network assesses the realism of the generated image compared to the ground-truth image . Pix2Pix has been applied to various I2I translation tasks, such as converting maps to satellite images, sketches to photos, and gray-scale images to color images. Figure 3 demonstrates the details of the Pix2Pix architecture with an example of converting sketches to real objects. The objective and loss function of Pix2Pix is similar to those of a CGAN, which can be written as:
(10) | ||||
Additionally, as noted in previous research [67], the generator can sometimes produce images that are too dissimilar from the target images, making it easy for the discriminator to identify them as fake. To address this issue, including an or loss term in the generator’s loss function is common. This loss term measures the difference between the generated image and the target image and penalizes the generator for producing images that are significantly different from the target. As such, an or loss term should be added to the generator itself. In Pix2Pix, loss is used to prevent the blurring of images:
(11) |
So that the final loss function to optimize the model is:
(12) |
U-Net generator: The Pix2Pix generator uses the U-Net architecture [71], which was state-of-the-art for image segmentation at the time of its development. The U-Net architecture consists of a series of down-sampling and up-sampling convolutional layers. The use of the U-Net architecture in the generator provides several benefits. One key advantage is its ability to handle large amounts of spatial information, making it well-suited for medical image segmentation tasks where the input images can be very large. Additionally, the symmetrical architecture of the U-Net enables the efficient transfer of information from the down-sampling path to the up-sampling path, preserving fine details of the segmentation mask. The U-Net also includes skip connections, which are implemented by concatenating activation from an earlier layer to the activation of a deeper layer. This allows the network to access high-level and low-level features, providing a complete representation of the input image. The skip connections also help capture long-range dependencies in the data by combining features from the contracting path with those from the expanding path.
PatchGAN discriminator: PatchGAN is a type of GAN discriminator that penalizes local image patch structures. It categorizes each patch in an image as either real or fake. The size of the patch can be adjusted as needed. One benefit of the PatchGAN discriminator is its ability to apply a fixed-size patch discriminator on images of any size, unlike typical discriminators that only score the full image. Additionally, a smaller PatchGAN has fewer parameters, runs faster, and can be applied to images of any size.
Training and optimization: GAN models are notoriously challenging to train and converge due to the simultaneous training of the generator and discriminator, with improvements to one model often coming at the expense of the other. The differing learning rates of the two models can also make convergence difficult. As suggested in the original GAN paper [23], instead of training the generator to minimize , it is more effective to train to maximize to improve convergence. In Pix2Pix, using instance normalization, which involves setting the batch size to 1, is a better choice than batch normalization for training generative models [87].
3.3 CycleGAN
General definition: CycleGAN [111] is another well-known I2I translation benchmark. It extends Pix2Pix by enabling unpaired data domain transmission, thereby resolving the issue of insufficient data. CycleGAN’s objective, given two distinct image domains and , is to learn the mapping between the two domains so that an image from domain can be transformed into an image in domain while preserving the original image’s content, and vice versa. It is used to enforce the cyclic consistency constraint between two image domains. If we translate an image from the domain to the domain and then back to the domain , we should obtain the original image. This constraint ensures that the generated images contain meaningful information and are not merely arbitrary noise. Figure 4 depicts the architecture of CycleGAN and its Cycle-consistency loss.

Loss function: There are two GANs in CycleGAN pipeline: represents the generator that map the image from domain to and the discriminator tries to identify the fake image and image from domain ; represents the generator that map the image from domain to and the discriminator tries to identify the fake image and image from domain . So there are two adversarial losses; for example, from the generator:
(13) | ||||
Meanwhile, the GAN that maps the image in domain to also has an objective function:
(14) | ||||
However, these two losses do not guarantee that the translated image will retain the domain-specific information. For instance, an image from domain that has been translated to domain may share similar characteristics with images in the domain while discarding all information from domain . CycleGAN uses a specific loss function to let the networks keep domain information, which is called cycle-consistency loss. The idea of forwarded cycle-consistency loss is that for each image from domain , the image translation cycle should be able to bring back to the original image, which means to minimize the distance between and . Similarly, the backward cycle-consistency loss is to minimize the distance between and . The cycle-consistency loss can be expressed by:
(15) | ||||
So that the total objective can be represented by:
(16) | ||||
4 MRI Reconstruction with Deep Learning
As stated in the prior chapter, I2I Translation has been utilized extensively in natural imaging. I2I also applies to medical imaging, including MRI and CT cross-domain translation. In contrast to I2I translation, MRI and multi-contrast MRI reconstruction attempts to recreate MR images using under-sampled k-space data rather than directly from other modalities.
Reference | Year | Dataset | Model |
---|---|---|---|
Non-generative model | |||
Wang et al[90] | 2016 | 500 fully sampled brain MRI | CNN |
Palangi et al[64] | 2016 | MNIST | LSTM+CS |
Yang et al[99] | 2018 | 100 fully sampled brain MR images | CNN+CS |
Schlemper[75] | 2017 | Fully sampled short-axis cardiac cine MR scans | CNN |
GAN-based model | |||
Yang et al[98] | 2017 | MICCAI 2013 grand challenge dataset | VGG-GAN |
Mardani et al[57] | 2019 | Knee dataset | |
Quan et al[69] | 2018 | MRI | GAN |
Jiang et al[35] | 2019 | MRI | WGAN |
Li et al[47] | 2019 | MICCAI 2013 grand challenge dataset | U-Net-GAN |
Murugesan et al.[60] | 2019 | MRIs | Recon-GLGAN |
Shaul et al[76] | 2020 | Multimodal MRI data | U-Net-GAN |
Chen et al[12] | 2021 | Multimodal MRI data | WPD-DAGAN |
Korkmaz et al[39] | 2021 | T1-weighted brain MRI | ViT-GAN |
Lin et al[48] | 2022 | knee and brain MRIs | ViT-GAN |
Diffusion model | |||
Cao et al[8] | 2022 | fastMRI dataset | |
Xie et al[97] | 2022 | fastMRI single-coil knee data | MC-DDPM |
4.1 Related Works
CNN is a potent DL model that has been widely applied to various computer vision problems ever since Alex et al. developed one of the most well-known CNN architectures, AlexNet [40]. CNN has evolved into a bench of permutations, such as VGGNet [79], ResNet [25], U-Net [71], etc. With the advent of GPUs, fairly deep CNNs such as ResNet have achieved remarkable success and can outperform most conventional techniques. However, CNNs are still considered black boxes at that time since they lack interpretation. Wang et al. [90] started to use CNN models instead of CS-MRI on the acceleration of MRI images by reducing the scanning time; they provide a supervised learning CNN model that can map the under-sampled images to fully-sampled images. Yang et al. [99] proposed a model-driven novel CNN model called ADMM-CSNet by reformulating the ADMM algorithms; it achieved better results than some data-driven CNN models.
After the CNN models become popular in the DL field, GANs can outperform conventional CNN models because they can learn data distributions significantly faster than CNN models. DAGAN [98] is the first well-known supervised GAN framework with conditions for MRI reconstruction. It employs a U-Net-like [71] architecture with the skip connection in the generator, a refined learning algorithm for fast convergence, and a novel combined loss function for superior reconstruction with preserved perceptual image details. Mardani et al. [57] provided a deep GAN to combine with the CS model, which can reduce the high-frequency noise and map zero-filled MR images to high-resolution MR images. Quan et al. [69] designed a dual-bench generator with ReconGAN and RefineGAN combining cyclic data consistency loss and generative adversarial loss, which can correctly map the under-sampled data to fully-reconstruction data. Li et al. [47] proposed the Structure-Enhanced GAN (SEGAN) with a new structure regularization called patch correlation regularization that aims to restore structure information locally and globally for efficient extraction of structure information. Murugesan et al. [60] proposed a novel GAN framework called reconstruction global-Local GAN (Recon-GLGAN), containing a generator and a context discriminator which incorporates global and local contextual information from images. Shaul et al. [77] introduced a practical, software-only GAN framework to estimate the missing k-space sample for brain MRI accelerating acquisition.
Most of the GAN models we discussed above use CNN-based generators and discriminators, especially U-Net and ResNet. They usually improve performance by redesigning loss functions or regularization. Recently, with the development of Vision Transformer (ViT), widely used in computer vision tasks that can outperform traditional CNN architectures, Some researchers have begun using ViTs as backbones instead of CNNs in GAN models. Korkmaz et al. [39] proposed a generative ViT called GVTrans, which can progressively map low-dimensional noise and latent variables onto MR images via cascaded blocks of cross-attention ViTs. Zhao et al. [109] demonstrated a GAN model called SwinGAN, which uses a Swin-Transformer-based backbone as the generator. Swin Transformer [50] is a U-Net shape hierarchical architecture ViT, which achieved robust results on various downstream tasks. Importantly, Lin et al. [48] given potential proof of the ViT-based GAN model can enable robust and fast accelerated MRI, using a pre-trained ViT model on natural images datasets, such as ImageNet, as the backbone of the GAN model that can be fine-tuned in MRI datasets which is costly to obtain. Table II shows some important GAN architectures for MRI reconstruction.
4.2 DAGAN
DAGAN [57] is one of the first de-aliasing and fast CS-MRI architectures that outperforms conventional CS-based methods and non-generative DL-based algorithms in MRI de-aliasing and reconstruction tasks. The DAGAN generator uses a CNN architecture similar to the U-Net [71] with skip connections, which is a highly advanced encoder-decoder network for medical image segmentation. DAGAN combines an adversarial loss with a novel content loss that considers both pixel-wise MSE and perceptual loss as defined by pre-trained VGGNet [79]. Additionally, frequency domain information from the CS-MRI is used as an additional data consistency constraint. Figure 5 illustrates the architecture of DAGAN.

In general, the DAGAN generator is based on the U-Net architecture and consists of 8 convolutional layers for down-sampling and eight corresponding convolutional layers for up-sampling. Each convolutional layer is followed by batch normalization and leaky ReLU layers, and there are skip connections between different levels of the encoder-decoder pairs. The discriminator performs a binary classification task between the generated de-aliased reconstruction image and the ground-truth image. The loss function of DAGAN is based on the GAN training process and can be expressed as:
(17) | |||
Where represents the ground-truth images from fully-sampled k-space and represents the zero-filled image. The generator produces the corresponding de-aliased reconstruction , which is then evaluated by the discriminator . A combined loss was designed for the generator’s training to improve the visual quality. This loss includes a pixel-wise mean square error (iMSE) loss, a frequency domain mean square error (fMSE) loss, and a perceptual VGG loss. The dual-MSE loss accounts for both the pixel-wise and frequency domain data, providing effective constraints, while the VGG loss enhances image quality. The two MSE losses for the image domain and frequency domain can be expressed as:
(18) | ||||
Where and are the corresponding frequency domain data of and after the Fourier transform. The VGG perceptual loss is represented by:
(19) |
So, the total combined loss can be written by:
(20) |
The DAGAN [57] loss function for its generator consists of a combination of adversarial loss, pixel-wise mean square error (iMSE) loss, frequency domain mean square error (fMSE) loss and a perceptual VGG loss [79]. This combination of losses allows DAGAN to overcome smooth reconstruction difficulties and gain more anatomical information than traditional CNN-based MRI reconstruction algorithms. The adversarial loss is denoted as , and and are hyper-parameters that control the weight of each loss. The dual-MSE and VGG loss have been widely adopted in many subsequent works [35, 102, 55, 5].
4.3 Recon/RefineGAN

Quan et al. [69] developed a highly regarded GAN-based MRI reconstruction framework, which blends a deep convolutional auto-encoder with the GAN architecture. This framework enables the utilization of deeper generator and discriminator networks to achieve improved performance. The framework is further enhanced by incorporating cyclic data consistency loss, which ensures faithful interpolation of under-sampled k-space data. Figure 6 shows the details of the Recon/RefineGAN framework.
The generator consists of two parts: ReconGAN and RefineGAN, each with an auto-encoder architecture with four encoder and decoder layers. Like DAGAN, Recon/RefineGAN proposed other loss constraints besides a general adversarial loss. They designed a cyclic data consistency loss , which is a combination of under-sampled frequency loss and fully reconstructed image loss in a cyclic fashion. The cyclic data consistency loss can be represented by:
(21) | ||||
Where is the observed under-sampled data from the frequency domain, represents the data in the frequency domain after the is taken into generator to get a reconstructed image, then apply under-sampled operator (Fourier transform and sampling mask) on it. Similarly, is the fully reconstructed image; if we use under-sampling operation on , then we get in the frequency domain that can be transformed into zero-filled MR image , the represents the image after is taken into generator . represents distance metrics, such as mean-square-error (MSE), and mean-absolute-error (MAE).
5 Multi-contrast MRI Reconstruction

In recent years, the development of various advanced GAN models has led to a surge in research on multi-contrast MRI reconstruction and synthesis. Medical image synthesis has been utilized to facilitate the conversion from one imaging modality to another in clinical settings, with the goal of reducing the amount of radiation exposure for patients. For example, mapping an MR image to a CT image can be considered cross-domain medical image synthesis [95, 94, 28]. This section will first provide an overview of general medical image synthesis and reconstruction problems and then delve into multiple recent GAN-based architectures that have shown excellent performance in MR image reconstructions.
Multi-contrast MRI is a diagnostic and therapeutic imaging technique that uses multiple imaging sequences within a single MRI scan to produce images with varying contrasts, providing a more in-depth understanding of the body’s structure and pathological changes. The imaging sequences, such as T1- and T2-weighted, proton-density-weighted, and contrast-enhanced MRI, each offer distinct information about the scanned tissue, such as the distinction between gray and white matter tissues in T1-weighted brain imaging and fluid and cortical tissue in T2-weighted brain imaging. Magnetic resonance angiography (MRA) is another crucial MR contrast that is used for imaging and assessing vascular anatomy and associated diseases. Different contrasts of MR images often contain similar information, making it possible to reconstruct between contrasts. This versatility of multi-contrast MRI makes it a powerful tool in the medical field [30].
5.1 Related Works
The first deep learning model for multi-contrast CS-MRI reconstruction with fewer model parameters was proposed by Sun et al. [83]. Olut et al. [62] developed a GAN-based model with a new loss term that accurately captures the directional features of vascular structures and can reconstruct MRA images from T1- and T2-weighted MRI images. Song et al. [81] utilized a coupled dictionary learning method to reconstruct T1- and T2-weighted MR images simultaneously. Wei et al. [93] created a dual-domain GAN for multi-contrast MRI reconstruction using two pairs of generators and discriminators in both the image and frequency (k-space) domains. Dar et al. [17] proposed a CGAN-based [59] generative model, which improved the quality and versatility of multi-contrast MRI exams compared to traditional nonlinear regression and deterministic neural networks. It successfully extends the natural image synthesis task into the medical imaging domain. Dar et al. [18] developed a new method that builds upon their previous work [17] and addresses the limitations of pure learning-based reconstruction or synthesis by incorporating three important priors. Yurt et al. [103] created a multi-stream GAN architecture for multi-contrast MRI synthesis called MustGAN, which built the adaptive fusion of unique features in multiple one-to-one streams and shared features in a many-to-one stream in order to learn more latent representations and improve the synthesis ability. Do et al. [19] proposed a dual-GAN-based network to jointly reconstruct T1- and T2-weighted MR images, with the optimization of sampling patterns and training processes verified.
5.2 cGAN for medical image synthesis
In Section II, we introduced the concept of conditional Generative Adversarial Network (cGAN). This technique is commonly used for image-to-image translation and synthesis tasks, such as converting grayscale images to color images and generating full images from partial images. In medical image synthesis, one of the most significant applications is the synthesis of T1- and T2-weighted MR images. Dar et al. [17] proposed a cGAN-based pipeline for the mutual translation of T1- and T2-weighted MR images. Instead of using the original adversarial loss function defined in equation (4), they utilized a squared adversarial loss function in their approach.
(22) | |||
Where represents the ground-truth image, represents the source image (T1- or T2-weighted) used as a condition, represents the random noise, and represents the reconstructed image. The cGAN has two branches of GAN pipelines that can translate between T1- and T2-weighted images.
5.3 Double-Domain GAN
In this section, we introduce a novel Double-Domain GAN (DDGAN) framework purposed by Wei et al. [93], which can reconstruct under-sampled T2-weight and MRA images from fully sampled T1-weighted images in a supervised manner. GAN has been applied in cross-modality image synthesis in recent years. In medical imaging, GAN-based image synthesis models can translate MRI, CT, and PET mutually. DDGAN is inspired by the possible combination of multi-contrast MRI reconstruction with synthesis. Specifically, DDGAN uses fully-sampled T1 k-space data as input with under-sampled k-space data from T2-weighted and MRA, which can share mutual information that improve the reconstruction quality of T2-weighted and MRA images. Moreover, DDGAN was created using a dual-domain GAN architecture, which consists of k-space domain GAN and image domain GAN. The k-space generator (K-generator) encapsulates the data using max-pooling and retains larger feature maps, which are more appropriate for data completion. A classical deep encoding network (U-net) is implemented for the image generator (I-generator) to extract multi-scale and dimensional features in image space for image denoising and enhancement [32, 80]. Together, these two GANs improved the efficacy of multi-contrast MRI reconstruction.
General definition: To illustrate it clearly, we can use the function to show the target for DDGAN:
(23) |
Where represents the loss function, and represent under-sampled k-space data of T2-weight and MRA images, while is the fully sampled k-space data of T1-weight images, and are fully sampled T2-weighted and MRA images in the image domain. The modal DDGAN can generate reconstructed T2-weighted and MRA images which should be indistinguishable from and from the image space discriminator. In general, the pipeline of DDGAN consists of two modules: the data completion module and the image enhancement module. Figure 7 shows details of DDGAN architecture.
Data completion module: In this module, a specific GAN with the generator and discriminator in the k-space domain are represented by and , respectively. Through training, is able to learn a mapping from under-sampled to fully-sampled k-space data, which is then distinguished from ground truth by . The input of includes the concatenation of fully-sampled k-space data in the T1-weighted image, under-sampled k-space data from the T2-weighted image, and MRA image. The output of includes a concatenation of fully-sampled k-space T1-weighted image, reconstructed k-space T2-weighted image, and MRA image, the should distinguish the output concatenation image from the ground-truth image (fully-sampled k-space T1-weighted image, T2-weighted image, and MRA image). is a data consistency operation that can replace the generated k-spaces with the data actually sampled. For GAN in the k-space domain, it provides a pixel-wised -loss between generated k-space pairs and standard k-space pairs. The -loss function is expressed by:
(24) | |||
Where, is the data consistency operation, is concatenation operation, is the loss. The adversarial loss function in the k-space domain can be represented by:
(25) | |||
Where is the given fully-sampled T1 image from k-space that can share information with other modalities. In general, the data completion module GAN can reconstruct k-space data from the under-sampled one, which can fill up the missing part of the k-space data.
Image enhancement module: In this module, the generator is taken inputs from an inverse Fourier Transform of the concatenation output of in the data completion module, including which means that can learn from fully-reconstructed k-space when the GAN in data complete module is finished training so that the GAN in image domain can focus on image enhancement and de-noising, instead of data translation from k-space. The output of can generate reconstructed T1-weighted MR images, T2-weighted MR images, and MRA images. The loss function in the image enhancement GAN can be represented by:
(26) | |||
Experiments and validation of DDGAN can be divided into three distinct levels. First, it demonstrated the efficacy of dual-GAN architecture by comparing the performance of the k-space generator to that of dual-GAN. (k-space and image-space GAN). Second, it indicates that the completely sampled T1-weighted image is useful for reconstructing T2-weighted and MRA images, which is superior to reconstructing T2-weighted and MRA images using their k-space data alone. Comparing DDGAN to other GAN architectures, such as Pix2Pix, and CS-based MRI reconstruction algorithms revealed that DDGAN outperformed all baselines. In Section VI, experimental details will be provided.
6 Experiments
In this section, we provide an overview of key evaluation metrics for I2I translation and other generative tasks. In addition, we present several essential evaluation metrics that could be used for MRI reconstruction. We identified significant I2I translation experiments, such as Pix2Pix, to demonstrate the benefits and drawbacks of its experimental outcomes and how we can enhance our research based on them. In addition to comparing I2I translation in natural imaging, we select a recent multi-contrast MRI reconstruction pipeline, DDGAN, to demonstrate its experimental results that may have a correlation with natural imaging, as well as to identify its potential advantages and disadvantages in order to direct future research.
6.1 Evaluation Metrics
For (I2I) translation tasks, several evaluation metrics are commonly used to measure the quality and similarity between the generated and ground truth images. Some examples of these metrics include:
Mean Absolute Error (MAE): MAE computes the average absolute difference between the image pixel values of the generated and ground truth images. It can be stated as follows:
(27) |
where is the number of samples, represents the predicted value for the sample and is the actual value. The absolute difference between these values is calculated using the absolute value operator .
Mean Squared Error (MSE): MSE is the average squared difference between the image pixel values generated and the ground truth. The MSE is formulated as follows:
(28) |
The MSE represents the average squared difference between predicted and actual values. In contrast to the MAE, the MSE emphasizes larger differences and is sensitive to outliers. However, MSE and MAE are pixel-level metrics that can only determine the pixel-wise distance; they are hard to measure the structural and textural similarity.
FCN score [52]: is an evaluation metric used to quantify the performance of a semantic segmentation model. In particular, people use FCN scores for tasks in I2I translation.
Structural Similarity Index (SSIM) [92]: SSIM takes into account changes in the mean, standard deviation, and covariance of the pixel values to determine the structural similarity between the generated and ground truth images. SSIM can be described as:
(29) |
The SSIM index evaluates the structural similarity between two images, and , by examining their mean intensity values ( and ), the variances of their intensity values ( and ), and the covariance between the intensity values of the two images (). To prevent instability in the division, and are utilized as small constants. The SSIM index ranges from -1 to 1, with a score closer to 1 indicating a high degree of structural similarity between the two images.
Peak Signal-to-Noise Ratio (PSNR): The PSNR is the ratio between the highest possible pixel value and the mean squared difference between the created image and the ground truth image. PSNR is represented by:
(30) |
Where MAX is the maximum possible pixel value, MSE is the Mean Squared Error between the predicted and ground truth image. To assess image reconstruction quality, the PSNR compares the maximum possible pixel value with the mean squared error between the predicted and ground truth images. The higher the PSNR, the lower the reconstruction error and the higher the quality of the reconstructed image.
Fréchet Inception Distance (FID): In MRI reconstruction, The FID is a widespread metric used to evaluate the quality of generated images compared to real images. It is defined as:
(31) |
Where and are probability distributions of real and generated images, respectively. and are the mean vectors of the real and generated image features, and and are their covariance matrices.
LPIPS (Learned Perceptual Image Patch Similarity) [107], which assesses the variety of the translated images and correlates well with a human perceptual similarity. It is calculated as the average LPIPS distance between two sets of translation outputs chosen randomly from the same input. A translated outcome that is more realistic and diverse will have a higher LPIPS score. The function of LPIPS can be expressed by:
(32) |
where is the LPIPS distance between two images and , and are the width and height of the images, respectively, and is the deep neural network used to extract features from the images. The term is the Euclidean distance between the feature vectors extracted from the corresponding pixels in the two images. The final result is the average distance over all the pixels in the images.
6.2 Datasets
Table 3 shows some important datasets that are used for I2I translation. Table 4 shows some datasets for MRI reconstruction, including T1, T2, PD, and MRA modalities for heart, knees, and brain research.
Dataset | Year | Total | Classes | Application |
---|---|---|---|---|
CelebA[51] | 2015 | 202,599 | 10177 | Facial attributes |
RaFD[41] | 2010 | 8040 | 67 | Facial expressions |
CMP Facades[86] | 2013 | 606 | 12 | Façade images |
Facescrub[61] | 2014 | 106,863 | 153 | Faces |
Cityscapes[16] | 2016 | 70,000 | 30 | Semantic |
Helen Face[42] | 2012 | 2330 | Face Parsing | |
CartoonSet[72] | 2018 | 10,000 | Cartoon Faces |
6.3 Experimental results
6.3.1 Pix2Pix
In Pix2Pix, the author attempts to discover which loss function is more significant using contrastive studies. As described in Section III, Pix2Pix features two loss functions: the adversarial loss, which is used to train the whole network, and the loss, which produces less blurred output than the loss. Table 5 shows that different loss functions could significantly impact generated images evaluated by FCN-Score [52]. The author also compared the U-Net generator to the conventional auto-encoder generator. Figure 8 indicates that U-net performs better at extracting and retaining high-level features such as textural information. Still, the auto-encoder lacks skip connection structures, and less information may transcend the bottleneck layer, resulting in several repeated lines on output images during upsampling procedures (Block A). In addition, U-net and auto-encoder generators cannot perform well only on loss, but U-Net is still better at capturing structural information (block B).
Loss | Per-pixel acc. | Per-class acc. | Class IOU |
---|---|---|---|
L1 | |||
GAN | |||
CGAN | |||
L1+GAN | |||
L1+CGAN | |||
Ground truth |

Furthermore, Pix2Pix has conducted a number of experiments, one of which compared the performance of a PatchGAN-like discriminator using various patch sizes. The results show that a patch size of performed better than a patch size of ; however, the patch size of performed worse than the patch size of . This experiment highlights the potential for decreased performance when using larger patch sizes for the PatchGAN discriminator. From my own perspective, I think PatchGAN has its own contribution to this pipeline so the experimental results to show its performance is important and necessary. Additionally, the author assessed the perceptual realism of colorization tasks by comparing the performance of different loss functions. The results showed that CGANs could produce compelling colorizations, but a common failure mode is the production of a gray-scale or desaturated result. These experimental results deliver a hint for future research directions that try to fix the problem related to gray-scale images. The author also evaluated the map vs. aerial photograph tasks using the Amazon Mechanical Turk (AMT) method, a perceptual evaluation by human discriminators. The results indicated that aerial photos generated by Pix2Pix could fool 18.9% of trials, which was significantly better than the baseline. However, maps generated by Pix2Pix only fooled 6.1% of trials, which could be due to structural errors in the generated maps that are more noticeable when compared to aerial photos. Pix2Pix has also been tested on semantic segmentation tasks, and CGAN with the model has achieved the best performance. In conclusion, when the Pix2Pix paper was published, although Pix2Pix has made a series of experiments, that showed a tremendous number of images as examples, there were fewer evaluation metrics, such as AMT and FCN-based evaluation, which may have defects in determining whether or not the results were satisfactory. This is the primary issue that the Pix2Pix examinations have uncovered.
6.3.2 DDGAN
DDGAN [93] is a typical multi-contrast MRI reconstruction work, which uses k-space and image domain dual-domain GAN to implement cross-contrast reconstruction for T2-weighted images and MRA by given fully-sampled T1-weighted images. In the experiment part, DDGAN uses the IXI dataset (http://braindevelopment.org/ixi-dataset/), which contains more than 600 MR images. They created 2D slices from T1-, T2- weighted, and MRA images and interpolate them into the same dimension and size, which is suitable for the network inputs. Finally, there are 4339 (80%) slices for training while 1101 (20%) slices for testing. Figure 9 [93] shows examples of image slices and under-sampling masks.

The author conducted multiple comparative experiments to demonstrate that DDGAN has exceeded its capacity. First, an experimental result demonstrates that DDGAN outperforms single-domain GAN, including K-space GAN and image-domain GAN, on all three evaluation metrics (PSNR, SSIM, RMSE) and various accelerating rates (). Second, to demonstrate the effectiveness of multi-input for reconstruction, the author built two models: one that directly reconstructs T2 and MRA images from their under-sampled k-space data, which is called rNet, and another that uses T1 contrast as source information to synthesize T2 and MRA images, which is called sNet. DDGAN performs substantially better than rNet and sNet. These results demonstrate that DDGAN employs completely sampled T1-weighted images as a single input that can share mutual information, which is advantageous when reconstructing other contrast images. Lastly, DDGAN compares its results to those of other models, such as CS-MRI-based traditional methods and the Pix2Pix-based model, which are used as benchmarks to demonstrate DDGAN’s success. Figure 10 shows the results comparison of different level baselines and DDGAN.

In conclusion, in my opinion, DDGAN has conducted a series of experiments that demonstrate the efficacy of its concept to combine the fully-sampled T1-weighted image as a successful image source. Furthermore, DDGAN has demonstrated that its performance can outperform some well-known architectures, such as Pix2Pix. When the undersampled mask pattern alters, DDGAN has also demonstrated its reconstruction details. However, DDGAN’s experiments are mainly focused on the performance of MRI reconstruction after the mutual information shared by multi-contrast training examples. However, it is hard to show that the model’s robustness, portability, and transferability. In the meantime, the GAN-based baselines of DDGAN are confined to Pix2Pix, and it is unknown how DDGAN compares to other pipelines that use multi-contrast MRI as research targets. DDGAN should, for instance, compare itself to Cycle-MedGAN [4]. In addition, DDGAN should attempt to demonstrate its ability to aid in clinical diagnosis by providing examples of its performance in actual diagnostic procedures.
7 Discussion
In our previous discussion, we touched upon the various interconnections and associations between natural image-to-image (I2I) translation and its implementation in the area of medical imaging. For instance, certain studies have adapted Pix2Pix for use in the medical field, effectively addressing medical image translation challenges [5]. Additionally, research utilizing CycleGAN in an unsupervised manner for medical image translation [4] demonstrates that cycle-consistency loss can be employed for image registration during bi-domain medical image translation. With the development of deep learning models in the field of natural imaging, there are an increasing number of works in the medical domain that use Pix2Pix or other well-known benchmarks as a starting point to enhance their work based on the prior state of the art. However, medical imaging has unique characteristics that may hinder the model’s performance in the natural image domain. For instance, some medical images have a distinct pattern that reveals information about specific disorders; therefore, our generative model must acquire this pattern effectively and precisely. In this section, we will discuss some significant challenges involving both the natural and medical fields, with the goal of defining a clearer and more distinct boundary that can guide our future research efforts.
7.1 Reviews in I2I translation
In the field of natural I2I translation, Pix2Pix has overcome several challenges and problems faced by researchers. Its most significant contribution is integrating separate, special-purpose machinery methods into a single framework that can solve different problems using a single pipeline [111]. Pix2Pix uses a U-Net instead of an auto-encoder as the generator’s backbone, which allows it to avoid losing structural information and preserve more details and textures. Additionally, Pix2Pix’s open-source software allows a wide range of tasks to be demonstrated without needing specific parameter tuning. Pix2Pix has become a successful benchmark in the I2I translation field, as evidenced by its use by various individuals, including artists, for their own purposes. However, Pix2Pix has limitations, such as the need for high-quality paired image data for training, which restricts the model’s ability to create diverse outputs. To address this, the same research group as Pix2Pix, Zhu et al., designed CycleGANs for I2I translation without using paired data as training samples, as the generators are trained using only unpaired data from both domains [111]. As a result, CycleGAN is a versatile tool for I2I translation when paired data is difficult or impossible to obtain. However, CycleGANs cannot fully bridge the gap between paired and unpaired data models, which means that the performance of models that use paired data is usually better than models using unpaired data. Researchers try to solve this by using weak or semi-supervised data, which could lead to a more effective solution, although labeling data is a time-consuming process. For instance, Park et al. achieved better results in I2I translation using unpaired data through contrastive learning and collaboration with the encoder-decoder, based on InfoNCE loss [63, 65].
Furthermore, Pix2Pix and its improved frameworks have limitations in generating multiple outputs or across multiple domains. To overcome this, Choi et al. introduced StarGAN v1 and v2 [13, 14], effectively addressing the multi-domain I2I translation problem. StarGAN uses a single pair of generators and discriminators to learn all mappings and generate multi-domain images. He et al. proposed the deliberation learning framework for I2I translation, which reviews and refines output images, building upon CycleGAN and StarGAN [26].
Preserving fine-grained details, such as texture and object boundaries, in the output image is often a challenge in I2I translation, and Pix2Pix also faces this limitation. To address this, HDPix2Pix was introduced, which can generate photo-realistic outputs by using a multi-scale generator network to handle complex structures in the output images [91]. Park et al. proposed the SPADE method, a conditional normalization method that provides better semantic information preservation than traditional normalization layers [66]. Lee et al. created DRIT++, a framework for disentangled representation that allows for preserving different image information and uses cross-cycle consistency loss for cyclic reconstruction training without paired data [43]. Shen et al. [78] claim that previous approaches, such as DRIT++ and Pix2PixHD, can handle complex structures in the output photos but may have serious limitations if the target domain images contain rich content and multiple disparate objects. They proposed an instance-aware approach that applies fine-grained local and global styles to the target picture spatially and improves the overall quality of the output through joint training.
In general, the challenges associated with I2I translation include the quality of the translation, the preservation of complex context and image structures, multi-domain I2I translation, and training with unpaired data. Researchers continue to enhance the performance of I2I translation tasks by refining algorithms and architectures, selecting suitable loss functions, integrating with sophisticated statistical approaches, and developing unique training pipelines.
7.2 Reviews of MRI reconstruction
Here, we highlight some of the key challenges researchers face in MRI reconstruction and suggest potential solutions based on a literature review. A major challenge is improving the perceptual quality of reconstructed images. Training networks solely on L1/L2 distance enforces overall image or patch similarity but fails to address local features such as anatomical sharpness, leading to blurring and fuzzy output. To address this, DAGAN [98] introduced new loss functions, such as the perceptual loss generated by a pre-trained VGG network, to extract high-level features from MRI images and enhance image quality and anatomical information. Chen et al. [10] proposed a novel encoder-decoder architecture that aims to increase reconstruction quality by preserving local features and sharpening anatomy by confining the GAN to follow the data distribution in the K-space, reducing the risk of artificial details. Li et al. [46] designed an edge-enhanced dual discriminator GAN that improves the reconstruction of edge details, preserving structural details and enhancing overall image quality. Higher-quality MRI reconstruction with fewer anti-aliasing artifacts may lead to more effective therapeutic applications.
Most MRI reconstruction algorithms reviewed rely on supervised learning and thus require fully-sampled ground-truth images, which can be time-consuming to acquire. One potential solution is to develop a semi-supervised or unsupervised learning model, although unsupervised learning often leads to subpar reconstruction quality. Bora et al. [7] proposed AmbientGAN, a statistical-based method for solving unsupervised image reconstruction using natural imaging datasets, opening up future MRI reconstruction research opportunities. Cole et al. [15] proposed a novel GAN method that quickly generates high-quality 3D MR images compared to baselines. When tested on a knee dataset, the gap between the proposed unsupervised GAN and the supervised GAN was relatively small, with a difference of 0.78% worse PSNR, 4.17% worse NRMSE, and equal SSIM. Efforts to reconstruct MR images without fully-sampled data may be faster and more accurate than conventional CS-MRI techniques. However, the state-of-the-art performance and results of supervised MRI reconstruction are still maintained.
Some studies have shifted their focus to image synthesis rather than reconstruction. For example, Sohail et al. [80] developed a GAN model to produce fake MR images across multiple modalities and contrasts. These fake images are newly generated signals rather than reconstructed ones, and thus the focus is solely on image synthesis and not reconstruction in the image domain. The ultimate goal of multi-contrast MRI reconstruction is to use the information from one fully-sampled contrast to reconstruct other MR modalities from the given sub-sampled k-space data. Similarly, [17] proposed a T1- and T2-weighted MR images mutual synthesis pipeline using only the image domain. Wei et al. [93] proposed a dual-domain GAN (DDGAN) that uses both fully sampled and subsampled k-space for optimal multi-contrast MRI reconstruction. Multiple contributions have been made by DDGAN, including the reconstruction of T2-weighted and MRA images using fully-sampled T1 and sub-sampled T2 and MRA k-space data. This demonstrates that thoroughly sampled T1-weighted k-space data can provide mutual information that facilitates the enhanced reconstruction of other modalities. Moreover, three-level experimental results indicate that DDGAN obtains the highest score across all evaluation metrics, demonstrating superior performance in recovering intricate structures and minimizing aliasing artifacts for accelerated multi-contrast MRI. But DDGAN has some problems, even though it is one of the newest and most advanced architectures for multi-contrast MRI reconstruction. The author has provided evaluation metrics that emphasize performance and precision, but they lack training performance. Was this DDGAN simple to train? Moreover, this architecture has incorporated fully sampled T1-weighted image pairs as a supervised term, which may be limited in terms of the data source. Finally, the author states that the I-generator requires larger feature maps for data completion, so using ViT-based architecture should be better than traditional CNNs because the self-attention layer in ViT makes it possible to embed information globally across the overall image [109, 20].
8 Future directions
GAN-based I2I translation and MRI reconstruction have numerous redesign opportunities. Improving the efficacy of I2I translation and MRI reconstruction can generally be accomplished in two significant ways: we can design novel regularization methods and loss functions to enhance the model’s portability and robustness, and we can change the architecture of GANs; for example, we can change the CNN-based generator and discriminator to ViTs. Second, we can abandon GAN-based generative models in favor of the diffusion model, which has recently gained popularity.
Currently, GAN models utilize various CNN architectures such as U-Net, ResNet, auto-encoders, etc., primarily focusing on convolutional layers for extracting low and high-level features from input images. However, Zhang et al. [105] introduced a self-attention GAN (SAGAN) that incorporates self-attention into the GAN framework. This is crucial in addressing long-range dependencies as it allows the model to consider relationships between all elements in the input sequence, regardless of their position. Recently, the ViT-based GAN architecture has gained popularity, as using self-attention mechanisms and Transformer models [89] has improved performance in various computer vision tasks. Lee et al. [44] took this a step further by using the ViT architecture as the backbone for both the generator and discriminator, paving the way for fully replacing CNN models in GAN frameworks with ViTs. However, the application of ViT-GAN models, which have primarily focused on natural image I2I translation and other generative tasks, has yet to be extended to the medical imaging domain. VTGAN [36] has made a step further toward medical image synthesis using ViT-based GAN architecture, it simultaneously performs semi-supervised image synthesis from fundus photographs to fluorescein angiography (FA) for the diagnosis of retinal disease.
Our review of related works on ViT-GAN models for MRI reconstruction and other medical imaging generative tasks [39] [48] showed improved performance compared to state-of-the-art CNN-based GAN models. Additionally, the recent use of ViT-GAN for self-supervised multi-contrast MRI reconstruction has opened up new research directions. Zhou et al. [110] proposed a novel self-supervised framework based on the swin-transformer, known as DSFormer, which leverages information sharing between different modalities to reconstruct MR images. This framework represents the most recent and innovative application of ViT-GAN for MRI reconstruction tasks. Moreover, we have lots of other future directions for GAN-based MRI reconstruction. Researchers may continue to develop more advanced GAN architectures that can handle more complex MRI data and provide higher-quality reconstructions. This may include incorporating additional loss functions or designing new training methods for robustness training.
On the other hand, the GAN-based generative model is not the only one that can be used to solve I2I translation and MRI reconstruction issues. The diffusion model, also known as denoising diffusion probabilistic models (DDPM) [29], is a novel Markov-chain-based generative model that can be utilized effectively for a variety of generative tasks, including I2I translation, generating high-resolution images and audio signals, etc. The fundamental concept underlying DDPMs is to characterize the noise distribution probabilistically. Given the observed data, the model is trained to understand the probability distribution of the noise. After learning this distribution, the model can be used to generate new images or audio signals by introducing noise and then removing it using the learned noise distribution. Sasaki et al. [74] purposed a novel generative model called UNIT-DDPM for unpaired I2I translation. It integrated the de-noising Markov Chain Monte Carlo approach to generate high-resolution outputs, which can outperform traditional GAN-based models by evaluation using novel FID metrics. In the medical imaging domain, diffusion-model-based generative methods are widely used for different objectives, one of the most interesting achievements, which is just accepted by CVPR 2023 is that Takagi et al. [84] is able to reconstruct a sequence of images that illustrate what the signal includes by applying the latent diffusion model to the visual fMRI signal; in other words, this work is able to materialize the images that the human brain recalled.
In addition, We can also apply interpretable deep learning techniques and models for MRI reconstruction and get a better interpretation and understanding of our models. The interpretable AI (XAI) has been widely used in the medical imaging domain [1], but has not been successfully used for MRI reconstruction tasks. Our research can still focus on developing real-time MRI reconstruction software to integrate our new clinical algorithms.
9 Conclusion
In our survey, we thoroughly examined various aspects of I2I translation in natural imaging and the challenges and premises of MRI reconstruction. Additionally, we explored the potential links between the natural image domain and the medical image domain in the context of generative deep learning techniques based on GANs. Our survey highlighted several key frameworks that have achieved state-of-the-art performance in the field of MRI reconstruction. We also identified emerging trends and potential directions for future research in this field.
References
- [1] S. H. P. Abeyagunasekera, Y. Perera, K. Chamara, U. Kaushalya, P. Sumathipala, and O. Senaweera. Lisa: Enhance the explainability of medical images unifying current xai techniques. In 2022 IEEE 7th International conference for Convergence in Technology (I2CT), pages 1–9. IEEE, 2022.
- [2] A. Almahairi, S. Rajeshwar, A. Sordoni, P. Bachman, and A. Courville. Augmented cyclegan: Learning many-to-many mappings from unpaired data. In International conference on machine learning, pages 195–204. PMLR, 2018.
- [3] M. Amodio and S. Krishnaswamy. Travelgan: Image-to-image translation by transformation vector learning. In Proceedings of the ieee/cvf conference on computer vision and pattern recognition, pages 8983–8992, 2019.
- [4] K. Armanious, C. Jiang, S. Abdulatif, T. Küstner, S. Gatidis, and B. Yang. Unsupervised medical image translation using cycle-medgan. In 2019 27th European signal processing conference (EUSIPCO), pages 1–5. IEEE, 2019.
- [5] K. Armanious, C. Jiang, M. Fischer, T. Küstner, T. Hepp, K. Nikolaou, S. Gatidis, and B. Yang. Medgan: Medical image translation using gans. Computerized medical imaging and graphics, 79:101684, 2020.
- [6] O. Bernard, A. Lalande, C. Zotti, F. Cervenansky, X. Yang, P.-A. Heng, I. Cetin, K. Lekadir, O. Camara, M. A. G. Ballester, et al. Deep learning techniques for automatic mri cardiac multi-structures segmentation and diagnosis: is the problem solved? IEEE transactions on medical imaging, 37(11):2514–2525, 2018.
- [7] A. Bora, E. Price, and A. G. Dimakis. Ambientgan: Generative models from lossy measurements. In International conference on learning representations, 2018.
- [8] C. Cao, Z.-X. Cui, S. Liu, D. Liang, and Y. Zhu. High-frequency space diffusion models for accelerated mri. arXiv preprint arXiv:2208.05481, 2022.
- [9] J. Cao, L. Hou, M.-H. Yang, R. He, and Z. Sun. Remix: Towards image-to-image translation with limited data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 15018–15027, 2021.
- [10] S. Chen, S. Sun, X. Huang, D. Shen, Q. Wang, and S. Liao. Data-consistency in latent space and online update strategy to guide gan for fast mri reconstruction. In Machine Learning for Medical Image Reconstruction: Third International Workshop, MLMIR 2020, Held in Conjunction with MICCAI 2020, Lima, Peru, October 8, 2020, Proceedings 3, pages 82–90. Springer, 2020.
- [11] X. Chen, C. Xu, X. Yang, and D. Tao. Attention-gan for object transfiguration in wild images. In Proceedings of the European conference on computer vision (ECCV), pages 164–180, 2018.
- [12] Y. Chen, D. Firmin, and G. Yang. Wavelet improved gan for mri reconstruction. In Medical Imaging 2021: Physics of Medical Imaging, volume 11595, pages 285–295. SPIE, 2021.
- [13] Y. Choi, M. Choi, M. Kim, J.-W. Ha, S. Kim, and J. Choo. Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 8789–8797, 2018.
- [14] Y. Choi, Y. Uh, J. Yoo, and J.-W. Ha. Stargan v2: Diverse image synthesis for multiple domains. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 8188–8197, 2020.
- [15] E. K. Cole, J. M. Pauly, S. S. Vasanawala, and F. Ong. Unsupervised mri reconstruction with generative adversarial networks. arXiv preprint arXiv:2008.13065, 2020.
- [16] M. Cordts, M. Omran, S. Ramos, T. Rehfeld, M. Enzweiler, R. Benenson, U. Franke, S. Roth, and B. Schiele. The cityscapes dataset for semantic urban scene understanding. In Proc. of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
- [17] S. U. Dar, M. Yurt, L. Karacan, A. Erdem, E. Erdem, and T. Cukur. Image synthesis in multi-contrast mri with conditional generative adversarial networks. IEEE transactions on medical imaging, 38(10):2375–2388, 2019.
- [18] S. U. Dar, M. Yurt, M. Shahdloo, M. E. Ildız, B. Tınaz, and T. Çukur. Prior-guided image reconstruction for accelerated multi-contrast mri via generative adversarial networks. IEEE Journal of Selected Topics in Signal Processing, 14(6):1072–1087, 2020.
- [19] W.-J. Do, S. Seo, Y. Han, J. C. Ye, S. H. Choi, and S.-H. Park. Reconstruction of multicontrast mr images through deep learning. Medical physics, 47(3):983–997, 2020.
- [20] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
- [21] H. Emami, M. M. Aliabadi, M. Dong, and R. B. Chinnam. Spa-gan: Spatial attention gan for image-to-image translation. IEEE Transactions on Multimedia, 23:391–401, 2020.
- [22] A. Gokaslan, V. Ramanujan, D. Ritchie, K. I. Kim, and J. Tompkin. Improving shape deformation in unsupervised image-to-image translation. In Proceedings of the European Conference on Computer Vision (ECCV), pages 649–665, 2018.
- [23] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. Generative adversarial networks. Communications of the ACM, 63(11):139–144, 2020.
- [24] M. A. Griswold, P. M. Jakob, R. M. Heidemann, M. Nittka, V. Jellus, J. Wang, B. Kiefer, and A. Haase. Generalized autocalibrating partially parallel acquisitions (grappa). Magnetic Resonance in Medicine: An Official Journal of the International Society for Magnetic Resonance in Medicine, 47(6):1202–1210, 2002.
- [25] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
- [26] T. He, Y. Xia, J. Lin, X. Tan, D. He, T. Qin, and Z. Chen. Deliberation learning for image-to-image translation. In IJCAI, pages 2484–2490, 2019.
- [27] J. Hennig. K-space sampling strategies. European radiology, 9(6):1020–1031, 1999.
- [28] Y. Hiasa, Y. Otake, M. Takao, T. Matsuoka, K. Takashima, A. Carass, J. L. Prince, N. Sugano, and Y. Sato. Cross-modality image synthesis from unpaired data using cyclegan: Effects of gradient consistency loss and training data size. In Simulation and Synthesis in Medical Imaging: Third International Workshop, SASHIMI 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, September 16, 2018, Proceedings 3, pages 31–41. Springer, 2018.
- [29] J. Ho, A. Jain, and P. Abbeel. Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33:6840–6851, 2020.
- [30] J. Huang, C. Chen, and L. Axel. Fast multi-contrast mri reconstruction. Magnetic resonance imaging, 32(10):1344–1352, 2014.
- [31] X. Huang, M.-Y. Liu, S. Belongie, and J. Kautz. Multimodal unsupervised image-to-image translation. In Proceedings of the European conference on computer vision (ECCV), pages 172–189, 2018.
- [32] C. M. Hyun, H. P. Kim, S. M. Lee, S. Lee, and J. K. Seo. Deep learning for undersampled mri reconstruction. Physics in Medicine & Biology, 63(13):135007, 2018.
- [33] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros. Image-to-image translation with conditional adversarial networks. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 5967–5976, 2017.
- [34] O. N. Jaspan, R. Fleysher, and M. L. Lipton. Compressed sensing mri: a review of the clinical literature. The British journal of radiology, 88(1056):20150487, 2015.
- [35] M. Jiang, Z. Yuan, X. Yang, J. Zhang, Y. Gong, L. Xia, and T. Li. Accelerating cs-mri reconstruction with fine-tuning wasserstein generative adversarial network. IEEE Access, 7:152347–152357, 2019.
- [36] S. A. Kamran, K. F. Hossain, A. Tavakkoli, S. L. Zuckerbrod, and S. A. Baker. Vtgan: Semi-supervised retinal image synthesis and disease prediction using vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3235–3245, 2021.
- [37] J. Kim, M. Kim, H. Kang, and K. Lee. U-gat-it: Unsupervised generative attentional networks with adaptive layer-instance normalization for image-to-image translation. arXiv preprint arXiv:1907.10830, 2019.
- [38] T. Kim, M. Cha, H. Kim, J. K. Lee, and J. Kim. Learning to discover cross-domain relations with generative adversarial networks. In International conference on machine learning, pages 1857–1865. PMLR, 2017.
- [39] Y. Korkmaz, M. Yurt, S. U. H. Dar, M. Özbey, and T. Cukur. Deep mri reconstruction with generative vision transformers. In International Workshop on Machine Learning for Medical Image Reconstruction, pages 54–64. Springer, 2021.
- [40] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural networks. In F. Pereira, C. Burges, L. Bottou, and K. Weinberger, editors, Advances in Neural Information Processing Systems, volume 25. Curran Associates, Inc., 2012.
- [41] O. Langner, R. Dotsch, G. Bijlstra, D. H. Wigboldus, S. T. Hawk, and A. Van Knippenberg. Presentation and validation of the radboud faces database. Cognition and emotion, 24(8):1377–1388, 2010.
- [42] V. Le, J. Brandt, Z. Lin, L. Bourdev, and T. S. Huang. Interactive facial feature localization. In Computer Vision–ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, October 7-13, 2012, Proceedings, Part III 12, pages 679–692. Springer, 2012.
- [43] H.-Y. Lee, H.-Y. Tseng, Q. Mao, J.-B. Huang, Y.-D. Lu, M. Singh, and M.-H. Yang. Drit++: Diverse image-to-image translation via disentangled representations. International Journal of Computer Vision, 128:2402–2417, 2020.
- [44] K. Lee, H. Chang, L. Jiang, H. Zhang, Z. Tu, and C. Liu. Vitgan: Training gans with vision transformers. arXiv preprint arXiv:2107.04589, 2021.
- [45] Q. Li, W. Cai, X. Wang, Y. Zhou, D. D. Feng, and M. Chen. Medical image classification with convolutional neural network. In 2014 13th international conference on control automation robotics & vision (ICARCV), pages 844–848. IEEE, 2014.
- [46] Y. Li, J. Li, F. Ma, S. Du, and Y. Liu. High quality and fast compressed sensing mri reconstruction via edge-enhanced dual discriminator generative adversarial network. Magnetic Resonance Imaging, 77:124–136, 2021.
- [47] Z. Li, T. Zhang, P. Wan, and D. Zhang. Segan: structure-enhanced generative adversarial network for compressed sensing mri reconstruction. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 1012–1019, 2019.
- [48] K. Lin and R. Heckel. Vision transformers enable fast and robust accelerated mri. In International Conference on Medical Imaging with Deep Learning, pages 774–795. PMLR, 2022.
- [49] M.-Y. Liu, T. Breuel, and J. Kautz. Unsupervised image-to-image translation networks. Advances in neural information processing systems, 30, 2017.
- [50] Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, and B. Guo. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 10012–10022, 2021.
- [51] Z. Liu, P. Luo, X. Wang, and X. Tang. Deep learning face attributes in the wild. In Proceedings of International Conference on Computer Vision (ICCV), December 2015.
- [52] J. Long, E. Shelhamer, and T. Darrell. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3431–3440, 2015.
- [53] F. Luan, S. Paris, E. Shechtman, and K. Bala. Deep photo style transfer. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4990–4998, 2017.
- [54] M. Lustig, D. L. Donoho, J. M. Santos, and J. M. Pauly. Compressed sensing mri. IEEE signal processing magazine, 25(2):72–82, 2008.
- [55] J. Lv, C. Wang, and G. Yang. Pic-gan: a parallel imaging coupled generative adversarial network for accelerated multi-channel mri reconstruction. Diagnostics, 11(1):61, 2021.
- [56] J. Lv, J. Zhu, and G. Yang. Which gan? a comparative study of generative adversarial network-based fast mri reconstruction. Philosophical Transactions of the Royal Society A, 379(2200):20200203, 2021.
- [57] M. Mardani, E. Gong, J. Y. Cheng, S. S. Vasanawala, G. Zaharchuk, L. Xing, and J. M. Pauly. Deep generative adversarial neural networks for compressive sensing mri. IEEE Transactions on Medical Imaging, 38(1):167–179, 2019.
- [58] A. M. Mendrik, K. L. Vincken, H. J. Kuijf, M. Breeuwer, W. H. Bouvy, J. De Bresser, A. Alansary, M. De Bruijne, A. Carass, A. El-Baz, et al. Mrbrains challenge: online evaluation framework for brain image segmentation in 3t mri scans. Computational intelligence and neuroscience, 2015:1–1, 2015.
- [59] M. Mirza and S. Osindero. Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784, 2014.
- [60] B. Murugesan, S. Vijaya Raghavan, K. Sarveswaran, K. Ram, and M. Sivaprakasam. Recon-glgan: A global-local context based generative adversarial network for mri reconstruction. In Machine Learning for Medical Image Reconstruction: Second International Workshop, MLMIR 2019, Held in Conjunction with MICCAI 2019, Shenzhen, China, October 17, 2019, Proceedings 2, pages 3–15. Springer, 2019.
- [61] H.-W. Ng and S. Winkler. A data-driven approach to cleaning large face datasets. In 2014 IEEE international conference on image processing (ICIP), pages 343–347. IEEE, 2014.
- [62] S. Olut, Y. H. Sahin, U. Demir, and G. Unal. Generative adversarial training for mra image synthesis using multi-contrast mri. In International workshop on predictive intelligence in medicine, pages 147–154. Springer, 2018.
- [63] A. v. d. Oord, Y. Li, and O. Vinyals. Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748, 2018.
- [64] H. Palangi, R. Ward, and L. Deng. Distributed compressive sensing: A deep learning approach. IEEE Transactions on Signal Processing, 64(17):4504–4518, 2016.
- [65] T. Park, A. A. Efros, R. Zhang, and J.-Y. Zhu. Contrastive learning for unpaired image-to-image translation. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part IX 16, pages 319–345. Springer, 2020.
- [66] T. Park, M.-Y. Liu, T.-C. Wang, and J.-Y. Zhu. Semantic image synthesis with spatially-adaptive normalization. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 2337–2346, 2019.
- [67] D. Pathak, P. Krahenbuhl, J. Donahue, T. Darrell, and A. A. Efros. Context encoders: Feature learning by inpainting. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2536–2544, 2016.
- [68] K. P. Pruessmann, M. Weiger, M. B. Scheidegger, and P. Boesiger. Sense: sensitivity encoding for fast mri. Magnetic Resonance in Medicine: An Official Journal of the International Society for Magnetic Resonance in Medicine, 42(5):952–962, 1999.
- [69] T. M. Quan, T. Nguyen-Duc, and W.-K. Jeong. Compressed sensing mri reconstruction using a generative adversarial network with a cyclic loss. IEEE transactions on medical imaging, 37(6):1488–1497, 2018.
- [70] S. Ren, K. He, R. Girshick, and J. Sun. Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in neural information processing systems, 28, 2015.
- [71] O. Ronneberger, P. Fischer, and T. Brox. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical image computing and computer-assisted intervention, pages 234–241. Springer, 2015.
- [72] A. Royer, K. Bousmalis, S. Gouws, F. Bertsch, I. Mosseri, F. Cole, and K. Murphy. Xgan: Unsupervised image-to-image translation for many-to-many mappings. In Domain Adaptation for Visual Understanding, pages 33–49. Springer, 2020.
- [73] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, et al. Imagenet large scale visual recognition challenge. International journal of computer vision, 115:211–252, 2015.
- [74] H. Sasaki, C. G. Willcocks, and T. P. Breckon. Unit-ddpm: Unpaired image translation with denoising diffusion probabilistic models. arXiv preprint arXiv:2104.05358, 2021.
- [75] J. Schlemper, J. Caballero, J. V. Hajnal, A. Price, and D. Rueckert. A deep cascade of convolutional neural networks for mr image reconstruction. In International conference on information processing in medical imaging, pages 647–658. Springer, 2017.
- [76] R. Shaul, I. David, O. Shitrit, and T. R. Raviv. Subsampled brain mri reconstruction by generative adversarial neural networks. Medical Image Analysis, 65:101747, 2020.
- [77] R. Shaul, I. David, O. Shitrit, and T. R. Raviv. Subsampled brain mri reconstruction by generative adversarial neural networks. Medical Image Analysis, 65:101747, 2020.
- [78] Z. Shen, M. Huang, J. Shi, X. Xue, and T. S. Huang. Towards instance-level image-to-image translation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 3683–3692, 2019.
- [79] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556, 2014.
- [80] M. Sohail, M. N. Riaz, J. Wu, C. Long, and S. Li. Unpaired multi-contrast mr image synthesis using generative adversarial networks. In Simulation and Synthesis in Medical Imaging: 4th International Workshop, SASHIMI 2019, Held in Conjunction with MICCAI 2019, Shenzhen, China, October 13, 2019, Proceedings, pages 22–31. Springer, 2019.
- [81] P. Song, L. Weizman, J. F. Mota, Y. C. Eldar, and M. R. Rodrigues. Coupled dictionary learning for multi-contrast mri reconstruction. IEEE transactions on medical imaging, 39(3):621–633, 2019.
- [82] P. L. Suárez, A. D. Sappa, and B. X. Vintimilla. Infrared image colorization based on a triplet dcgan architecture. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pages 18–23, 2017.
- [83] L. Sun, Z. Fan, X. Fu, Y. Huang, X. Ding, and J. Paisley. A deep information sharing network for multi-contrast compressed sensing mri reconstruction. IEEE Transactions on Image Processing, 28(12):6141–6153, 2019.
- [84] Y. Takagi and S. Nishimoto. High-resolution image reconstruction with latent diffusion models from human brain activity. bioRxiv, pages 2022–11, 2022.
- [85] H. Tang, D. Xu, N. Sebe, Y. Wang, J. J. Corso, and Y. Yan. Multi-channel attention selection gan with cascaded semantic guidance for cross-view image translation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 2417–2426, 2019.
- [86] R. Tyleček and R. Šára. Spatial pattern templates for recognition of objects with regular structure. In Proc. GCPR, Saarbrucken, Germany, 2013.
- [87] D. Ulyanov, A. Vedaldi, and V. Lempitsky. Instance normalization: The missing ingredient for fast stylization. arXiv preprint arXiv:1607.08022, 2016.
- [88] H. Uzunova, J. Ehrhardt, and H. Handels. Memory-efficient gan-based domain translation of high resolution 3d medical images. Computerized Medical Imaging and Graphics, 86:101801, 2020.
- [89] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. u. Kaiser, and I. Polosukhin. Attention is all you need. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30, 2017.
- [90] S. Wang, Z. Su, L. Ying, X. Peng, S. Zhu, F. Liang, D. Feng, and D. Liang. Accelerating magnetic resonance imaging via deep learning. In 2016 IEEE 13th international symposium on biomedical imaging (ISBI), pages 514–517. IEEE, 2016.
- [91] T.-C. Wang, M.-Y. Liu, J.-Y. Zhu, A. Tao, J. Kautz, and B. Catanzaro. High-resolution image synthesis and semantic manipulation with conditional gans. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 8798–8807, 2018.
- [92] Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli. Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing, 13(4):600–612, 2004.
- [93] H. Wei, Z. Li, S. Wang, and R. Li. Undersampled multi-contrast mri reconstruction based on double-domain generative adversarial network. IEEE Journal of Biomedical and Health Informatics, 2022.
- [94] J. M. Wolterink, A. M. Dinkla, M. Savenije, P. R. Seevinck, C. van den Berg, and I. Išgum. Mr-to-ct synthesis using cycle-consistent generative adversarial networks. Proc. Neural Inf. Process. Syst.(NIPS), 2017.
- [95] J. M. Wolterink, A. M. Dinkla, M. H. Savenije, P. R. Seevinck, C. A. van den Berg, and I. Išgum. Deep mr to ct synthesis using unpaired data. In Simulation and Synthesis in Medical Imaging: Second International Workshop, SASHIMI 2017, Held in Conjunction with MICCAI 2017, Québec City, QC, Canada, September 10, 2017, Proceedings 2, pages 14–23. Springer, 2017.
- [96] W. Wu, K. Cao, C. Li, C. Qian, and C. C. Loy. Transgaga: Geometry-aware unsupervised image-to-image translation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 8012–8021, 2019.
- [97] Y. Xie and Q. Li. Measurement-conditioned denoising diffusion probabilistic model for under-sampled medical image reconstruction. arXiv preprint arXiv:2203.03623, 2022.
- [98] G. Yang, S. Yu, H. Dong, G. Slabaugh, P. L. Dragotti, X. Ye, F. Liu, S. Arridge, J. Keegan, Y. Guo, et al. Dagan: deep de-aliasing generative adversarial networks for fast compressed sensing mri reconstruction. IEEE transactions on medical imaging, 37(6):1310–1321, 2017.
- [99] Y. Yang, J. Sun, H. Li, and Z. Xu. Admm-csnet: A deep learning approach for image compressive sensing. IEEE transactions on pattern analysis and machine intelligence, 42(3):521–538, 2018.
- [100] Z. Yi, H. Zhang, P. Tan, and M. Gong. Dualgan: Unsupervised dual learning for image-to-image translation. In Proceedings of the IEEE international conference on computer vision, pages 2849–2857, 2017.
- [101] Y. Yuan, S. Liu, J. Zhang, Y. Zhang, C. Dong, and L. Lin. Unsupervised image super-resolution using cycle-in-cycle generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pages 701–710, 2018.
- [102] Z. Yuan, M. Jiang, Y. Wang, B. Wei, Y. Li, P. Wang, W. Menpes-Smith, Z. Niu, and G. Yang. Sara-gan: Self-attention and relative average discriminator based generative adversarial networks for fast compressed sensing mri reconstruction. Frontiers in Neuroinformatics, 14:611666, 2020.
- [103] M. Yurt, S. U. Dar, A. Erdem, E. Erdem, K. K. Oguz, and T. Çukur. mustgan: multi-stream generative adversarial networks for mr image synthesis. Medical image analysis, 70:101944, 2021.
- [104] J. Zbontar, F. Knoll, A. Sriram, T. Murrell, Z. Huang, M. J. Muckley, A. Defazio, R. Stern, P. Johnson, M. Bruno, et al. fastmri: An open dataset and benchmarks for accelerated mri. arXiv preprint arXiv:1811.08839, 2018.
- [105] H. Zhang, I. Goodfellow, D. Metaxas, and A. Odena. Self-attention generative adversarial networks. In International conference on machine learning, pages 7354–7363. PMLR, 2019.
- [106] R. Zhang, P. Isola, and A. A. Efros. Colorful image colorization. In Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part III 14, pages 649–666. Springer, 2016.
- [107] R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O. Wang. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 586–595, 2018.
- [108] R. Zhang, J.-Y. Zhu, P. Isola, X. Geng, A. S. Lin, T. Yu, and A. A. Efros. Real-time user-guided image colorization with learned deep priors. arXiv preprint arXiv:1705.02999, 2017.
- [109] X. Zhao, T. Yang, B. Li, and X. Zhang. Swingan: A dual-domain swin transformer-based generative adversarial network for mri reconstruction. Computers in Biology and Medicine, 153:106513, 2023.
- [110] B. Zhou, N. Dey, J. Schlemper, S. S. M. Salehi, C. Liu, J. S. Duncan, and M. Sofka. Dsformer: A dual-domain self-supervised transformer for accelerated multi-contrast mri reconstruction. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), pages 4966–4975, January 2023.
- [111] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE international conference on computer vision, pages 2223–2232, 2017.
- [112] X. Zhu, L. Zhang, L. Zhang, X. Liu, Y. Shen, and S. Zhao. Gan-based image super-resolution with a novel quality loss. Mathematical Problems in Engineering, 2020:1–12, 2020.