Memory-efficient Learning for High-Dimensional MRI Reconstruction
Abstract
Deep learning (DL) based unrolled reconstructions have shown state-of-the-art performance for under-sampled magnetic resonance imaging (MRI). Similar to compressed sensing, DL can leverage high-dimensional data (e.g. 3D, 2D+time, 3D+time) to further improve performance. However, network size and depth are currently limited by the GPU memory required for backpropagation. Here we use a memory-efficient learning (MEL) framework which favorably trades off storage with a manageable increase in computation during training. Using MEL with multi-dimensional data, we demonstrate improved image reconstruction performance for in-vivo 3D MRI and 2D+time cardiac cine MRI. MEL uses far less GPU memory while marginally increasing the training time, which enables new applications of DL to high-dimensional MRI.
Keywords:
Magnetic Resonance Imaging (MRI) Unrolled reconstruction Memory-efficient learning.1 Introduction
Deep learning-based unrolled reconstructions (Unrolled DL recons)[2, 19, 1, 5, 21, 11] have shown great success at under-sampled MRI reconstruction, well beyond the capabilities of parallel imaging and compressed sensing (PICS)[12, 4, 16]. These methods are often formulated by unrolling the iterations of an image reconstruction optimization[5, 1, 21] and use a training set to learn an implicit regularization term represented by a deep neural network. It has been shown that increasing the number of unrolls improves upon finer spatial and temporal textures in the reconstruction[5, 1, 17]. Similar to compressed sensing and other low-dimensional representations, DL recons can take advantage of additional structure in very high-dimensional data (e.g. 3D, 2D+time, 3D+time) to further improve image quality. However, these large-scale DL recons are currently limited by GPU memory required for gradient-based optimization using backpropagation. Therefore, most Unrolled DL recons focus on 2D applications or are limited to a small number of unrolls. In this work, we use our recently proposed memory-efficient learning (MEL) framework[9, 23] to reduce the memory needed for backpropagation, which enables the training of Unrolled DL recons for 1) larger-scale 3D MRI; and 2) 2D+time cardiac cine MRI with a large number of unrolls (Figure 1). We evaluate the spatio-temporal complexity of our proposed method on the Model-based Deep Learning (MoDL) architecture [1] and train these high-dimensional DL recons on a single 12GB GPU. Our training uses far less memory while only marginally increasing the computation time. To demonstrate the advantages of high-dimensional reconstructions to image quality, we performed experiments on both retrospectively and prospectively under-sampled data for 3D MRI and cardiac cine MRI. Our in-vivo experiments indicate that by exploiting high-dimensional data redundancy, we can achieve better quantitative metrics and improved image quality with sharper edges for both 3D MRI and cardiac cine MRI.

2 Methods
2.1 Memory-efficient learning
As shown in Figure 2 a), unrolled DL recons are often formulated by unrolling the iterations of an image reconstruction optimization[5, 1]. Each unroll consists of two submodules: CNN based regularization layer and data consistency (DC) layer. In conventional backpropagation, the gradient must be computed for the entire computational graph, and intermediate variables from all unrolls need to be stored at a significant memory cost. By leveraging MEL, we can process the full graph as a series of smaller sequential graphs. As shown in Figure 2 b), first, we forward propagate the network to get the output without computing the gradients. Then, we rely on the invertibility of each layer (required) to recompute each smaller auto-differentiation (AD) graph from the network’s output in reverse order. MEL only requires a single layer to be stored in memory at a time, which reduces the required memory by a factor of . Notably, the required additional computation to invert each layer only marginally increases the backpropagation runtime.

2.2 Memory-efficient learning for MoDL
Here, we use a widely used Unrolled DL Recon framework: MoDL[1]. We formulate the reconstruction of as an optimization problem and solve it as below:
(1) |
where is the system encoding matrix, denotes the k-space measurements and is a learned CNN-based denoiser. For multi-channel MRI reconstruction, can be formulated as , where represent the multi-channel sensitivity maps, denotes Fourier Transform and is the undersampling mask used for selecting the acquired data. MoDL solves the minimization problem by an alternating procedure:
(2) |
(3) |
which represents the CNN-based regularization layer and DC layer respectively. In this formulation, the DC layer is solved using Conjugate Gradient (CG)[20], which is unrolled for a finite number of iterations. For all the experiments, we used an invertible residual convolutional neural network (RCNN) introduced in [3, 14, 6], whose architecture is composed of a 5-layer CNN with 64 channels per layer. Detailed network architecture is shown in Figure S1. The residual CNN is inverted using the fixed-point algorithm as described in [9], while the DC layer is inverted through:
(4) |
2.3 Training and evaluation of memory-efficient learning
With IRB approval and informed consent/assent, we trained and evaluated MEL on both retrospective and prospective 3D knee and 2D+time cardiac cine MRI. We conducted 3D MoDL experiments with and without MEL on 20 fully-sampled 3D knee datasets (320 slices each) from mridata.org[18]. 16 cases were used for training, 2 cases were used for validation and other 2 for testing. Around 5000 3D slabs with size 21256320 were used for training the reconstruction networks. All data were acquired on a 3T GE Discovery MR 750 with an 8-channel HD knee coil. An 8x Poisson Disk sampling pattern was used to retrospectively undersample the fully sampled k-space. Scan parameters included a matrix size of 320256320, and TE/TR of 25ms/1550ms. In order to further demonstrate the feasibility of our 3D reconstruction with MEL on realistic prospectively undersampled scans, we reconstructed 8 prospectively undersampled 3D FSE knee scans (available at mridata.org) with the model trained on retrospectively undersampled knee data. Scanning parameters includes: Volume size: 320288236, TR/TE = 1400/20.46ms, Flip Angle: 90∘, FOV: 160 mm160 mm 141.6 mm.
For the cardiac cine MRI, fully-sampled bSSFP cardiac cine datasets were acquired from 15 volunteers at different cardiac views and slice locations on 1.5T and 3.0T GE scanners using a 32-channel cardiac coil. All data were coil compressed[24] to 8 virtual coils. Twelve of the datasets (around 190 slices) were used for training, 2 for validation, and one for testing. k-Space data were retrospectively under-sampled using a variable-density k-t sampling pattern to simulate 14-fold acceleration with 25% partial echo. We also conducted experiments on a prospectively under-sampled scan (R=12) which was acquired from a pediatric patient within a single breath-hold on a 1.5T scanner.
We compared the spatio-temporal complexity (GPU memory, training time) with and without MEL. In order to show the benefits of high-dimensional DL recons, we compared the reconstruction results of PICS, 2D and 3D MoDL with MEL for 3D MRI, and 2D+time MoDL with 4 unrolls and 10 unrolls for cardiac cine MRI. For both 2D MoDL and 3D MoDL with MEL, we used 5 unrolls, 10 CG steps and Residual CNN as the regularization layer. A baseline PICS reconstruction was performed using BART[22]. Sensitivity maps were computed using BART[22] and SigPy[13]. Common image quality metrics such as Peak Signal to Noise Ratio (pSNR), Structual Similarity (SSIM) [8] and Fréchet Inception Distance (FID)[7] were reported. FID is a widely used measure of perceptual similarity between two sets of images. All the experiments were implemented in Pytorch [15] and used Nvidia Titan XP (12GB) and Titan V CEO (32GB) GPUs. Networks were trained end-to-end using a per-pixel loss and optimized using Adam [10] with a learning rate of .
3 Results
We first evaluate the spatio-temporal complexity of MoDL with and without MEL (Figure 3). Without MEL, for a 12GB GPU memory limit, the maximum slab size decreases rapidly as the number of unrolls increases, which limits the performance of a 3D reconstruction. In contrast, using MEL, the maximum slab size is roughly constant. Figure 3 b) and c) show the comparisons from two different perspectives: 1)GPU memory usage; 2)Training time per epoch. Results indicate that for both 3D and 2D+time MoDL, MEL uses significantly less GPU memory than conventional backpropagation while marginally increasing training time. Notably, both MoDL with and without MEL have the same inference time.

Figure 4 shows a comparison of different methods for 3D reconstruction. Instead of learning from only 2D axial view slices (Figure 1 a), 3D MoDL with MEL captures the image features from all three dimensions. Zoomed-in details indicate that 3D MoDL with MEL is able to provide more faithful contrast with more continuous and realistic textures as well as higher pSNR over other methods. Figure 5 demonstrates that MEL enables the training of 2D+time MoDL with a large number of unrolls (10 unrolls), which outperforms MoDL with 4 unrolls with respect to image quality and y-t motion profile. With MEL, MoDL with 10 unrolls resolves the papillary muscles (yellow arrows) better than MoDL with 4 unrolls. Also, the y-t profile of MoDL with 10 unrolls depicts motion in a more natural way while MoDL with 4 unrolls suffers from blurring. Meanwhile, using 10 unrolls over 4 unrolls yields an improvement of 0.6dB in validation pSNR.


Table 1 shows the quantitative metric comparisons (pSNR, SSIM and FID) between different methods on both 3D MRI and cardiac cine MRI reconstructions. The results indicate that both 3D MoDL with MEL and 2D+time MoDL with MEL outperform other methods with respect to pSNR, SSIM and FID.
metric | method | 3D MRI | 2D cardiac cine MRI |
---|---|---|---|
pSNR (dB) | PICS | 31.011.97 | 24.692.74 |
2D MoDL | 31.442.07 | - | |
3D MoDL with MEL | 32.112.05 | - | |
2D+time MoDL: 4 unrolls | - | 26.872.98 | |
2D+time MoDL with MEL: 10 unrolls | - | 27.423.21 | |
SSIM | PICS | 0.8160.046 | 0.8240.071 |
2D MoDL | 0.8210.044 | - | |
3D MoDL with MEL | 0.8300.038 | - | |
2D+time MoDL: 4 unrolls | - | 0.8700.042 | |
2D+time MoDL with MEL: 10 unrolls | - | 0.8880.042 | |
FID | PICS | 46.71 | 39.40 |
2D MoDL | 43.58 | - | |
3D MoDL with MEL | 41.48 | - | |
2D+time MoDL: 4 unrolls | - | 36.93 | |
2D+time MoDL with MEL: 10 unrolls | - | 31.64 |
Figure 6 a) and Figure S2 show the reconstruction results on two representative prospectively undersampled 3D FSE knee scan. Note that in this scenario, there is no fully-sampled groud truth. Despite there exists some difference between the training and testing (e.g., matrix size, scanning parameters), 3D MoDL with MEL is still able to resolve more detailed texture and sharper edges over traditional PICS and learning-based 2D MoDL. Figure 6 b) and Video S1 shows the reconstruction on a representative prospective undersampled cardiac cine scan. We can clearly see that enabled by MEL, 2D+time MoDL with 10 unrolls can better depicts the finer details as well as more natural motion profile.

4 Conclusions
In this work, we show that MEL enables learning for high-dimensional MR reconstructions on a single 12GB GPU, which is not possible with standard backpropagation methods. We demonstrate MEL on two representative large-scale MR reconstruction problems: 3D volumetric MRI, 2D cardiac cine MRI with a relatively large number of unrolls. By leveraging the high-dimensional image redundancy and a large number of unrolls, we were able to get improved quantitative metrics and reconstruct finer details, sharper edges, and more continuous textures with higher overall image quality for both 3D and 2D cardiac cine MRI. Furthermore, 3D MoDL reconstruction results from prospectively undersampled k-space show that the proposed method is robust to the scanning parameters and could be potentially deployed in clinical systems. Overall, MEL brings a practical tool for training the large-scale high-dimensional MRI reconstructions with much less GPU memory and is able to achieve improved reconstructed image quality.
5 Acknowledgements
The authors would like to thank Dr. Gopal Nataraj for his helpful discusses and paper editing. We also acknowledge support from NIH R01EB009690, NIH R01HL136965, NIH R01EB026136 and GE Healthcare.
References
- [1] Aggarwal, H.K., Mani, M.P., Jacob, M.: Modl: Model-based deep learning architecture for inverse problems. IEEE transactions on medical imaging 38(2), 394–405 (2018)
- [2] Diamond, S., Sitzmann, V., Heide, F., Wetzstein, G.: Unrolled optimization with deep priors. arXiv preprint arXiv:1705.08041 (2017)
- [3] Gomez, A.N., Ren, M., Urtasun, R., Grosse, R.B.: The reversible residual network: Backpropagation without storing activations. arXiv preprint arXiv:1707.04585 (2017)
- [4] Griswold, M.A., Jakob, P.M., Heidemann, R.M., Nittka, M., Jellus, V., Wang, J., Kiefer, B., Haase, A.: Generalized autocalibrating partially parallel acquisitions (grappa). Magnetic Resonance in Medicine: An Official Journal of the International Society for Magnetic Resonance in Medicine 47(6), 1202–1210 (2002)
- [5] Hammernik, K., Klatzer, T., Kobler, E., Recht, M.P., Sodickson, D.K., Pock, T., Knoll, F.: Learning a variational network for reconstruction of accelerated mri data. Magnetic resonance in medicine 79(6), 3055–3071 (2018)
- [6] He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 770–778 (2016)
- [7] Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. arXiv preprint arXiv:1706.08500 (2017)
- [8] Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 1125–1134 (2017)
- [9] Kellman, M., Zhang, K., Markley, E., Tamir, J., Bostan, E., Lustig, M., Waller, L.: Memory-efficient learning for large-scale computational imaging. IEEE Transactions on Computational Imaging 6, 1403–1414 (2020)
- [10] Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)
- [11] Küstner, T., Fuin, N., Hammernik, K., Bustin, A., Qi, H., Hajhosseiny, R., Masci, P.G., Neji, R., Rueckert, D., Botnar, R.M., et al.: Cinenet: deep learning-based 3d cardiac cine mri reconstruction with multi-coil complex-valued 4d spatio-temporal convolutions. Scientific reports 10(1), 1–13 (2020)
- [12] Lustig, M., Donoho, D., Pauly, J.M.: Sparse mri: The application of compressed sensing for rapid mr imaging. Magnetic Resonance in Medicine: An Official Journal of the International Society for Magnetic Resonance in Medicine 58(6), 1182–1195 (2007)
- [13] Ong, F., Lustig, M.: Sigpy: a python package for high performance iterative reconstruction. In: Proc. ISMRM (2019)
- [14] Ovadia, Y., Fertig, E., Ren, J., Nado, Z., Sculley, D., Nowozin, S., Dillon, J.V., Lakshminarayanan, B., Snoek, J.: Can you trust your model’s uncertainty? evaluating predictive uncertainty under dataset shift. arXiv preprint arXiv:1906.02530 (2019)
- [15] Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., Desmaison, A., Kopf, A., Yang, E., DeVito, Z., Raison, M., Tejani, A., Chilamkurthy, S., Steiner, B., Fang, L., Bai, J., Chintala, S.: Pytorch: An imperative style, high-performance deep learning library. In: Wallach, H., Larochelle, H., Beygelzimer, A., d'Alché-Buc, F., Fox, E., Garnett, R. (eds.) Advances in Neural Information Processing Systems 32, pp. 8024–8035. Curran Associates, Inc. (2019), http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf
- [16] Pruessmann, K.P., Weiger, M., Scheidegger, M.B., Boesiger, P.: Sense: sensitivity encoding for fast mri. Magnetic Resonance in Medicine: An Official Journal of the International Society for Magnetic Resonance in Medicine 42(5), 952–962 (1999)
- [17] Sandino, C.M., Lai, P., Vasanawala, S.S., Cheng, J.Y.: Accelerating cardiac cine mri using a deep learning-based espirit reconstruction. Magnetic Resonance in Medicine 85(1), 152–167 (2021)
- [18] Sawyer, A.M., Lustig, M., Alley, M., Uecker, P., Virtue, P., Lai, P., Vasanawala, S.: Creation of fully sampled mr data repository for compressed sensing of the knee. Citeseer (2013)
- [19] Schlemper, J., Caballero, J., Hajnal, J.V., Price, A.N., Rueckert, D.: A deep cascade of convolutional neural networks for dynamic mr image reconstruction. IEEE transactions on Medical Imaging 37(2), 491–503 (2017)
- [20] Shewchuk, J.R., et al.: An introduction to the conjugate gradient method without the agonizing pain (1994)
- [21] Tamir, J.I., Yu, S.X., Lustig, M.: Unsupervised deep basis pursuit: Learning inverse problems without ground-truth data. arXiv preprint arXiv:1910.13110 (2019)
- [22] Uecker, M., Ong, F., Tamir, J.I., Bahri, D., Virtue, P., Cheng, J.Y., Zhang, T., Lustig, M.: Berkeley advanced reconstruction toolbox. In: Proc. Intl. Soc. Mag. Reson. Med. No. 2486 in 23 (2015)
- [23] Zhang, K., Kellman, M., Tamir, J.I., Lustig, M., Waller, L.: Memory-efficient learning for unrolled 3d mri reconstructions. In: ISMRM Workshop on Data Sampling and Image Reconstruction (2020)
- [24] Zhang, T., Pauly, J.M., Vasanawala, S.S., Lustig, M.: Coil compression for accelerated imaging with cartesian sampling. Magnetic resonance in medicine 69(2), 571–582 (2013)
6 Supplementary Material


