3452

Learning Deep Linear Convolutional Transforms For Accelerated MRI
Hongyi Gu1,2, Burhaneddin Yaman1,2, Steen Moeller2, Il Yong Chun3, and Mehmet Akçakaya1,2
1Electrical and Computer Engineering, University of Minnesota, Minneapolis, MN, United States, 2Center for Magnetic Resonance Research, University of Minnesota, Minneapolis, MN, United States, 3Electrical and Computer Engineering, University of Hawai’i at Mānoa, Honolulu, HI, United States

Synopsis

Research shows that deep learning (DL) based MRI reconstruction outperform conventional methods, such as parallel imaging and compressed sensing (CS). Unlike CS with pre-determined linear representations for regularization, DL uses nonlinear representations learned from a large database. Transform learning (TL) is another line of work bridging the gap between these two approaches. In this work, we combine ideas from CS, TL and DL to learn deep linear convolutional transforms, which has comparable performance to DL and supports uniform under-sampling unlike CS, while enabling sparse convex optimization at inference time.

INTRODUCTION

Recently, many studies have shown that deep learning (DL), in particular physics-guided DL (PG-DL) methods have improved performance compared to parallel imaging1 and compressed sensing (CS)2 for accelerated MRI3, 4. CS uses a pre-specified linear representation of images for regularization, while PG-DL utilizes sophisticated non-linear representations implicitly learned through neural networks. Transfer learning (TL) is another line of work bridging the gap between these two approaches by learning linear representation from data5. DL methods use advanced optimization methods, and a large number of parameters, in contrast to the hand-tuning of two or three parameters in CS, or the more traditional optimization strategies employed in TL. On the other hand, sparse processing of linear representations used in CS/TL may provide a clearer understanding of robustness/stability6. Recent literature has aimed to address this gap by incorporating modern data science tools for improving linear representations of MR images, including denoising7 and revisiting conventional $$$\ell_1$$$-wavelet CS using state-of-the-art data science tools8. In this work, we combine ideas from CS, TL and DL for deep linear convolutional transform learning (DLC-TL). Results show that the performance gap between the proposed model and PG-DL is minimal; while our method reliably reconstructs datasets with uniform undersampling, unlike conventional CS.

MATERIALS AND METHODS

Inverse Problem for Accelerated MRI: The inverse problem for accelerated MRI is given as
$$\arg \min _{\mathbf{x}} \frac{1}{2}\|\mathbf{y}-\mathbf{E x}\|_{2}^{2}+\mathcal{R}(\mathbf{x})\tag{1}$$
where $$$\mathbf{y}$$$ is multi-coil k-space, $$$\mathbf{E}$$$ is forward multi-coil encoding operator1, $$$\|\mathbf{y}-\mathbf{E x}\|_{2}^{2}$$$ enforces data consistency (DC) and $$$\mathcal{R}(.)$$$ is a regularizer. In CS, $$$\mathcal{R}(\mathbf{x})$$$ is the weighted $$$\ell_1$$$-norm of transform coefficients in a pre-specified domain, such as wavelets2. In TL, $$$\mathcal{R}(\mathbf{x})$$$ uses a linear transform pre-learned from the dataset, or learned at the same time with the reconstruction5, while enforcing non-degenerate solutions with scaling constraints, or tight frame conditions5,9. This optimization is coupled over $$$\mathbf{x}$$$ and the linear transform, and run until convergence. In PG-DL, an optimization algorithm for solving (1) is unrolled for a fixed number of iterations. The regularizer in PG-DL is implemented implicitly via neural networks, while DC unit is solved linearly. The network is then trained end-to-end with appropriate loss functions, such as $$$\ell_2$$$ or $$$\ell_1$$$ norms.

Proposed Physics-Guided Deep Linear Convolutional Transform Learning: We set $$$\mathcal{R}(\mathbf{x})=\sum_{\mathbf{n}=1}^{N} \boldsymbol{\mu}_{\mathbf{n}}\left\|\mathbf{W}_{\mathbf{n}} \mathbf{x}\right\|_{1}$$$ in (1), where $$$\mathbf{W}_\mathbf{n}$$$ are linear convolutional transforms9, and solve the objective function via unrolled ADMM for $$$T$$$ iterations (Fig. 1), leading to tunable parameters for $$$\ell_1$$$ soft-thresholding, augmented Lagrangian relaxation and dual update. These parameters and $$$\mathbf{W}_\mathbf{n}$$$ are trained end-to-end, similar to PG-DL. We encode each $$$\ell_1$$$ as a deep linear network10, featuring cascades of convolutional layers, each designed to have distinct receptive field sizes. The cascades of convolutions leverage the optimization landscape of large-scale optimization algorithms used in DL with multiple local minima11, enabling the presence of infinitely many good-performing solutions for the training objective and faster convergence12. We refer to this overall approach as deep linear convolutional transform learning (DLC-TL).

For the proposed DLC-TL, we used $$$N$$$ = 6 linear transforms and $$$T$$$ = 10 unrolls. Table 1 summarizes the network structures being implemented for each $$$\mathbf{W}_\mathbf{n}$$$. Adam optimizer with learning rate 5×10-3 was used for training over 100 epochs, with a mixed normalized $$$\ell_1$$$-$$$\ell_2$$$ loss that uses a regularizer to enforce a pseudo-tight frame condition9 on $$$\mathbf{W}_\mathbf{n}$$$. DC sub-problem for ADMM was solved using CG13 with 5 iterations and warm-start.

Imaging Data: Fully-sampled coronal proton density (PD) with and without fat-suppression (PD-FS) knee data were obtained from the NYU-fast MRI database14. The datasets were retrospectively under-sampled with a uniform mask (R=4, 24 ACS lines). Training was performed on 300 slices from 10 subjects. Testing was performed on 10 different subjects. The proposed approach was compared with PG-DL based on same ADMM unrolling that utilized a ResNet-based regularizer unit15,16. Note the only difference between the methods is the $$$\mathcal{R}(.)$$$ term, where our approach uses learnable deep convolutional operations for solving a convex problem, while the other uses a CNN for implicit regularization. An $$$\ell_1$$$-Wavelet CS approach was also used for comparison.

RESULTS

Fig. 2 and 3 show representative reconstructions from coronal PD and PD-FS knee MRI, respectively. The proposed DLC-TL and PG-DL remove aliasing artifacts that are present in $$$\ell_1$$$-Wavelet CS. As CS reconstruction typically works with random undersampling patterns, its performance is degraded with the uniform undersampling in this work. Sharpness is maintained by DLC-TL and PG-DL, whereas blurring is more apparent with $$$\ell_1$$$-Wavelet CS, especially for coronal PD-FS data. Fig. 4 depicts the quantitative metrics over all test datasets, showing the median and interquartile ranges (25th-75th percentile) of NMSE and SSIM metrics. Both the proposed DLC-TL and PG-DL outperform $$$\ell_1$$$-Wavelet CS, while having comparable quantitative metrics.

DISCUSSION AND CONCLUSIONS

In this study, we proposed a combination of ideas from TL, CS and PG-DL literatures to learn deep linear convolutional transform for MRI reconstruction. While PG-DL quantitatively outperforms our approach, the performance gap is minimal. Our proposed method leads to a linear representation, and enables convex optimization at inference time.

Acknowledgements

Grant support: NIH R01HL153146, NIH P41EB027061, NIH U01EB025144; NSF CAREER CCF-1651825

References

1.Pruessmann K P, Weiger M, Bornert P, Boesiger P. Advances in sensitivity encoding with arbitrary k-space trajectories. Magn Reson Med. 2001;46(4):638-651.

2.Lustig M, Donoho D, Pauly J. Sparse MRI: The application of compressed sensing for rapid MR imaging. Magn Reson Med. 2007;58(6):1182-1195.

3.Knoll F, Hammernik K, Zhang C, et al. Deep-learning methods for parallel magnetic resonance imaging reconstruction. IEEE Sig Proc Mag. 2020;37(1):128-140.

4.Liang D, Cheng J, Ke Z, Ying L. Deep magnetic resonance image reconstruction: inverse problems meet neural networks. IEEE Sig Proc Mag. 2020;37(1):141-151.

5.Wen B, Ravishankar S, Pfister L, Bresler Y. Transform learning for magnetic resonance image reconstruction: From model-based learning to building neural networks. IEEE Sig Proc Mag. 2020;37(1):41-53.

6.Ye J C. Compressed sensing MRI: a review from signal processing perspective. BMC Biomed Eng. 2019;1:8.

7.Ramzi Z, Starck J, Moreau T, Ciuciu P. Wavelets in the deep learning era. in Proc. IEEE EUSIPCO. 2021;1417-1421.

8.Gu H, Yaman B, Uğurbil K, et al. Compressed sensing MRI with $$$\ell_1$$$-wavelet reconstruction revisited using modern data science tools. in Proc. IEEE EMBC. 2021.

9.Chun I Y, Fessler J A. Convolutional analysis operator learning: Acceleration and convergence. IEEE Trans Image Process. 2020;29:2108-2122.

10.Saxe A M, McClelland J L, Ganguli S. Exact solutions to the nonlinear dynamics of learning in deep linear neural networks. arXiv preprint. 2014;arXiv:1312.6120.

11.Choromanska A, Henaff M, Mathieu M, et al. The loss surfaces of multilayer networks. J Magn Reason. 2015;38:192-204.

12.Bell-Kligler S, Shocher A, Irani M. Blind super-resolution kernel estimation using an internal-GAN. in Proc. NeurIPS. 2019.

13.Aggarwal H K, Mani M P, Jacob M. MoDL: Model-based deep learning architecture for inverse problems. IEEE Trans Med Imaging. 2019;38(2):394-405.

14.Knoll F, Zbontar J, Sriram A, et al. fastMRI: A publicly avail-able raw k-space and DICOM dataset of knee images for accelerated MR image reconstruction using machine learning. Radiol AI. 2020;2(1):e190007.

15.Hosseini S A H, Yaman B, Moeller S, et al. Dense recurrent neural networks for accelerated MRI: History-cognizant unrolling of optimization algorithms. IEEE J Sel Top Signal Process. 2020;14(6):1280-1291.

16.Yaman B, Hosseini S A H, Moeller S, Ellermann J, Uğurbil K, Akçakaya M. Self-supervised learning of physics-guided reconstruction neural networks without fully sampled reference data. Magn Reson Med. 2020;84:3172-3191.

Figures

Figure 1. Schematic of the unrolled ADMM for physics-guided reconstruction with deep linear convolutional transform learning (DLC-TL). An unrolled iteration of ADMM consists of a regularizer (R) with linear transforms, data consistency (DC) and dual update. All learnable parameters are shared across unrolled iterations. Each $$$\mathbf{W}_\mathbf{n}$$$ is encoded as a deep linear network with cascades of convolutional (Conv) layers.

Figure 2. A representative slice from coronal PD knee MRI, reconstructed using $$$\ell_1$$$-Wavelet CS, proposed DLC-TL, and PG-DL. Unlike $$$\ell_1$$$-Wavelet CS, the latter two methods have low level of residual artifacts (yellow arrows), and the proposed DLC-TL performs closely to PG-DL.

Figure 3. A representative coronal PD-FS knee slice, reconstructed using $$$\ell_1$$$-Wavelet CS, proposed DLC-TL, and PG-DL. The latter two methods have significantly less fewer artifacts than $$$\ell_1$$$-Wavelet CS (yellow arrows), and the proposed DLC-TL method has comparable performance and sharpness to PG-DL.

Figure 4. Boxplots showing the median and interquartile range (25th-75th percentile) of (a) NMSE and (b) SSIM metrics on test slices from 10 subjects for coronal PD and PD-FS datasets. Both the proposed DLC-TL and PG-DL outperform $$$\ell_1$$$-Wavelet CS, while having comparable quantitative metrics.

Table 1. Specifications of each linear convolutional transform.

Proc. Intl. Soc. Mag. Reson. Med. 30 (2022)
3452
DOI: https://doi.org/10.58530/2022/3452