3287

Alternating Joint Learning Approach for Variational Networks and Sampling Pattern in Parallel MRI
Marcelo Victor Wust Zibetti1, Florian Knoll2, and Ravinder Regatte1
1Radiology, NYU Langone Health, New York, NY, United States, 2Department of Artificial Intelligence in Biomedical Engineering, FAU Erlangen-Nuremberg, Erlangen, Germany

Synopsis

We propose a new alternating learning approach to jointly learn the sampling pattern (SP) and the parameters of a variational network (VN) for acquisition and reconstruction on 3D Cartesian parallel MRI problems. This approach is composed of alternating short training with BASS algorithm to learn the SP, and ADAM algorithm to learn the parameters of the VN, both with forced monotonicity. The results illustrate that this approach provides reduced error when compared to other joint learning approaches, and surpasses VN trained with recently developed fixed SPs.

Introduction:

Undersampling the k-space has been successfully used to accelerate MRI data acquisition, particularly when combined with parallel MRI (1,2), and special reconstructions (3–6). Compressive sensing (CS) has shown that undersampling artifacts from incoherent acquisition can be recovered with sparsity-enforced reconstructions (7). Recently, deep learning image reconstructions have shown that neural networks (NN), such as the variational networks (VN) (8), can also effectively remove artifacts from undersampled images (5,8,9). However, no specifications for the sampling pattern (SP) are currently known to be more or less effective with NN. Due to this, several approaches have been proposed to jointly learn the SP and the NN (10–12), as an attempt to understand what are the sampling requirements for effective reconstruction with NN.
Here, we propose a new alternating approach to learn the SP and the parameters of a VN. The approach considers SP learning as a combinatorial problem while keeping the VN learning as regular NN training. However, forced monotonicity is used to ensure that the cost function is being reduced. Our approach shows that the learned SP and VN can surpass other jointly learned approaches, such as LOUPE (10), as well as VN learning with the most recent SPs for CS approaches, such as the Poisson-disc with variable density (PD+VD) SP (13,14).

Methods:

In the VN for undersampled parallel MRI we use the following model:
$$\mathbf{m}=\mathbf{FC}\mathbf{x}.$$
In the above, $$$\mathbf{x}$$$ represents the 2D+time images, of size $$$N_x\times N_y \times N_t$$$ which denotes vertical $$$N_x$$$ and horizontal $$$N_y$$$ sizes and time $$$N_t$$$. $$$\mathbf{m}$$$ is the fully-sampled multi-coil k-t-space data. $$$\mathbf{C}$$$ denotes the coil sensitivities transform, which maps $$$\mathbf{x}$$$ into multi-coil weighted images of size $$$N_x \times N_y \times N_t \times N_c$$$, with number of coils $$$N_c$$$. $$$\mathbf{F}$$$ represents the spatial FFTs, which are $$$N_t \times N_c$$$ repetitions of the 2D-FFT.
When undersampled is used, then:
$$\bar{\mathbf{m}}=\mathbf{S}_Ω\mathbf{FC}\mathbf{x},$$
where, $$$\mathbf{S}_Ω $$$ is the sampling function using SP $$$Ω$$$ (same for all coils). The SP contains $$$M$$$ the k-t-space positions that will be sampled from a total of $$$N=N_x \times N_y \times N_t$$$ possible positions. The acceleration factor (AF) is defined as $$$N/M$$$.
In a VN (8), the reconstruction is computed by a neural network with $$$J$$$ layers that are inspired in gradient descent algorithms:
$$\mathbf{x}_{j+1} = \mathbf{x}_{j} –(\alpha_j \mathbf{C}^* \mathbf{F}^* \mathbf{S}^*_Ω(\bar{\mathbf{m}}- \mathbf{S}_Ω\mathbf{FC}\mathbf{x}_j ) + \sum_{f=1}^{N_f} \mathbf{K}^b_{j,f} \mathbf{\phi}’(\mathbf{K}_{j,f} \mathbf{x}_j )),$$
where $$$1 \leq j \leq J+1$$$ represents the layer index, where $$$J=10$$$ was used in this work.
All the VN parameters, i.e. convolutional filters $$$\mathbf{K}^b_{j,f}, \mathbf{K}_{j,f}$$$ ($$$N_f=24$$$, with complex-valued spatio-temporal filters of size $$$11×11×N_t$$$) and step-sizes $$$\alpha_j$$$, are learned from data. In this study, activation functions $$$\mathbf{\phi}’$$$ are fixed rectified linear units (ReLu).
The output of the VN can be written as:
$$\hat{\mathbf{x}}=R_{\theta}(\bar{\mathbf{m}}, Ω),$$
where $$$R_{\theta}$$$ represents the VN with parameters $$$\theta = \left\{ \left\{ \mathbf{K}^b_{j,f}, \mathbf{K}_{j,f},\right\}_{f=1}^{N_f} , \alpha_j \right\}_{j=1}^J$$$.
Our proposed approach, illustrated in Fig. 1 formulates the joint learning problem as an alternating minimization problem:
$$ Ω_{m+1} =\arg\min_{ \begin{array}{c}Ω \subset \Gamma\\ s.t. |Ω|=M\end{array}} \frac{1}{N_i} \sum_{i=1}^{N_i} f(\mathbf{x}_i , R_{\theta_m}( \mathbf{S}_Ω\mathbf{FC}\mathbf{x}_i, Ω).$$
$$ \theta_{m+1} = \arg\min_{ \theta \in \Theta} \frac{1}{N_i} \sum_{i=1}^{N_i} f(\mathbf{x}_i , R_{\theta}( \mathbf{S}_{Ω_{m+1}} \mathbf{FC}\mathbf{x}_i, Ω_{m+1})).$$
In the equations above, $$$ N_i$$$ is the number of images used for training. To learn the SP, some iterations of BASS (15) are used ($$$K_{init}=512$$$, $$$\alpha=0.5$$$, and stops when $$$K=1$$$), while to learn the VN parameters, some iterations of ADAM (16) are used ($$$8$$$ epochs with initial learning-rate of $$$2\times10^{-4}$$$, with a learning-rate drop factor of $$$0.25$$$, applied every $$$2$$$ epochs, and batch size of $$$8$$$ images). Monotonicity is forced in both algorithms.
We compared the proposed approach against LOUPE (10), modified for parallel MRI (17), and a VN using a fixed PD+VD SP (13). All learning approaches started with pre-trained VN ($$$80$$$ epochs, ADAM algorithm, with an initial learning-rate of $$$2\times10^{-4}$$$, learning-rate drop factor of $$$0.5$$$ applied every $$$5$$$ epochs, and batch size of $$$8$$$ images), trained with diverse SPs and diverse AFs. Because LOUPE learns a sampling density (SD) instead of an SP, we retrained the VN with one randomly generated SP from the learned LOUPE SD.
We assessed the root mean squared error (RMSE) on two datasets: brain and knee. The brain dataset contains $$$260$$$ images of size $$$N=320 \times 320 \times 1$$$ for training and $$$60$$$ for testing. The knee dataset contains the same number of images of size $$$N=128 \times 64 \times 10$$$ that is used for T1ρ mapping (18,19).

Results and Discussion:

In Fig. 2 the resulting RMSE for both datasets at different AF is shown. Note that the proposed approach achieved better RMSE than the other approaches (between $$$14.9%\~51.2%$$$). A comparison between the learned SPs is shown in Fig. 3. Note that the proposed approach learned the low and mid frequencies are more important to be sampled.
In Fig. 4, some visual examples with the brain dataset are shown. Note that the proposed approach seems less noisy and recovered some structures better. In Fig. 5, some results for the knee dataset are shown. In this example, the ability to learn temporal sampling is also shown.

Conclusion

Our alternating learning approach showed improvements in accelerated MRI when compared to other joint learning approaches or VN trained with fixed SPs.

Acknowledgements

This study was supported by NIH grants, R21-AR075259-01A1, R01-AR068966, R01-AR076328-01A1, R01-AR076985-01A1, and R01-AR078308-01A1 and was performed under the rubric of the Center of Advanced Imaging Innovation and Research (CAI2R), an NIBIB Biomedical Technology Resource Center (NIH P41-EB017183).

References

1. Ying L, Liang Z-P. Parallel MRI Using Phased Array Coils. IEEE Signal Process. Mag. 2010;27:90–98 doi: 10.1109/MSP.2010.936731.

2. Blaimer M, Breuer F, Mueller M, Heidemann RM, Griswold MA, Jakob PM. SMASH, SENSE, PILS, GRAPPA. Top. Magn. Reson. Imaging 2004;15:223–236 doi: 10.1097/01.rmr.0000136558.09801.dd.

3. Tamir JI, Ong F, Anand S, Karasan E, Wang K, Lustig M. Computational MRI With Physics-Based Constraints: Application to Multicontrast and Quantitative Imaging. IEEE Signal Process. Mag. 2020;37:94–104 doi: 10.1109/MSP.2019.2940062.

4. Feng L, Benkert T, Block KT, Sodickson DK, Otazo R, Chandarana H. Compressed sensing for body MRI. J. Magn. Reson. Imaging 2017;45:966–987 doi: 10.1002/jmri.25547.

5. Knoll F, Hammernik K, Zhang C, et al. Deep-learning methods for parallel magnetic resonance imaging reconstruction: A survey of the current approaches, trends, and issues. IEEE Signal Process. Mag. 2020;37:128–140 doi: 10.1109/MSP.2019.2950640.

6. Jacob M, Ye JC, Ying L, Doneva M. Computational MRI: Compressive Sensing and Beyond [From the Guest Editors]. IEEE Signal Process. Mag. 2020;37:21–23 doi: 10.1109/MSP.2019.2953993.

7. Lustig M, Donoho DL, Santos JM, Pauly JM. Compressed sensing MRI. IEEE Signal Process. Mag. 2008;25:72–82 doi: 10.1109/MSP.2007.914728.

8. Hammernik K, Klatzer T, Kobler E, et al. Learning a variational network for reconstruction of accelerated MRI data. Magn. Reson. Med. 2018;79:3055–3071 doi: 10.1002/mrm.26977.

9. Liang D, Cheng J, Ke Z, Ying L. Deep magnetic resonance image reconstruction: Inverse problems meet neural networks. IEEE Signal Process. Mag. 2020;37:141–151 doi: 10.1109/MSP.2019.2950557.

10. Bahadir CD, Wang AQ, Dalca A V., Sabuncu MR. Deep-Learning-Based Optimization of the Under-Sampling Pattern in MRI. IEEE Trans. Comput. Imaging 2020;6:1139–1152 doi: 10.1109/TCI.2020.3006727.

11. Aggarwal HK, Jacob M. J-MoDL: Joint Model-Based Deep Learning for Optimized Sampling and Reconstruction. IEEE J. Sel. Top. Signal Process. 2020;14:1151–1162 doi: 10.1109/JSTSP.2020.3004094.

12. Weiss T, Vedula S, Senouf O, Michailovich O, Zibulevsky M, Bronstein A. Joint Learning of Cartesian under Sampling and Reconstruction for Accelerated MRI. In: IEEE International Conference on Acoustics, Speech and Signal Processing. Vol. 2020-May. IEEE; 2020. pp. 8653–8657. doi: 10.1109/ICASSP40776.2020.9054542.

13. Levine E, Daniel B, Vasanawala S, Hargreaves B, Saranathan M. 3D Cartesian MRI with compressed sensing and variable view sharing using complementary poisson‐disc sampling. Magn. Reson. Med. 2017;77:1774–1785 doi: 10.1002/mrm.26254.

14. Dwork N, Baron CA, Johnson EMI, O’Connor D, Pauly JM, Larson PEZ. Fast variable density Poisson-disc sample generation with directional variation for compressed sensing in MRI. Magn. Reson. Imaging 2021;77:186–193 doi: 10.1016/j.mri.2020.11.012.

15. Zibetti MVW, Herman GT, Regatte RR. Fast data-driven learning of parallel MRI sampling patterns for large scale problems. Sci. Rep. 2021;11:19312 doi: 10.1038/s41598-021-97995-w.

16. Kingma DP, Ba J. Adam: A Method for Stochastic Optimization. arXiv Prepr. 2014:1–15.

17. Zhang J, Zhang H, Wang A, et al. Extending LOUPE for K-Space Under-Sampling Pattern Optimization in Multi-coil MRI. Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics) 2020;12450 LNCS:91–101 doi: 10.1007/978-3-030-61598-7_9.

18. Zibetti MVW, Sharafi A, Otazo R, Regatte RR. Accelerating 3D-T1ρ mapping of cartilage using compressed sensing with different sparse and low rank models. Magn. Reson. Med. 2018;80:1475–1491 doi: 10.1002/mrm.27138.

19. Zibetti MVW, Johnson PM, Sharafi A, Hammernik K, Knoll F, Regatte RR. Rapid mono and biexponential 3D-T1ρ mapping of knee cartilage using variational networks. Sci. Rep. 2020;10:19144 doi: 10.1038/s41598-020-76126-x.

Figures

Fig. 1: General overview of the proposed approach for jointly learn the SP and the parameters of a VN on Parallel MRI problems.

Fig. 2: RMSE of the compared approaches: the initial pre-trained VN with VD+PD SP, a re-trained VN with one VD+PD SP, the VN trained with LOUPE, a re-trained VN with one SP from LOUPE, and the proposed approach. In (a) the results for the brain dataset are shown, and in (b) the results for the knee joint dataset are shown.

Fig.3: Comparing SPs for different methods and AFs for the brain dataset. In (a)-(c) the SPs learned by the proposed approach, for AF=3, 6, and 12, are shown. In (d)-(f) the VD+PD SPs are shown, for the same AFs. In (g)-(i) the SPs generated from the sampling densities (SD) learned in LOUPE, that are shown in (j)-(l). Yellow areas correspond to fixed auto-calibration areas of the SP.

Fig. 4: Comparing SP and images with brain dataset at AF=20. In (a)-(b) original reference image and its details are shown. In (c)-(f), the SP, VN reconstruction, error map, and details are shown for re-trained VN with LOUPE (RMSE=0.77). In (g)-(j), the same information for re-trained VN with VD+PD SP are shown (RMSE=0.70), and in (k)-(n) for the proposed approach (RMSE=0.55). Arrows in the detailed images point to the most relevant differences.

Fig. 5: In (a) the original image, used as a reference, is shown. In (b)-(d), the SP, VN reconstruction, and error map are shown for re-trained VN with LOUPE (RMSE=0.72). In (e)-(g), the same information for re-trained VN with VD+PD SP are shown (RMSE=0.62), and in (h)-(j) for the proposed approach (RMSE=0.56).

Proc. Intl. Soc. Mag. Reson. Med. 30 (2022)
3287
DOI: https://doi.org/10.58530/2022/3287