3614

Model-augmented deep learning for VFA-T1 mapping
Lea Bogensperger1,2, Oliver Maier1, and Rudolf Stollberger1,3
1Institute of Medical Engineering, Graz University of Technology, Graz, Austria, 2Institute for Computer Graphics and Vision, Graz University of Technology, Graz, Austria, 3Biotechmed, Graz, Austria

Synopsis

A deep learning approach is proposed to estimate M0 and T1 maps from undersampled variable flip angle (VFA) data to explore the potential of this method for acceleration and rapid reconstruction even without parallel imaging. A U-Net was implemented with a model consistency term containing the signal equation to ensure the physical validity and to include prior knowledge of B1+. Training is performed on numerical brain phantoms and by means of transfer-learning on retrospectively undersampled in-vivo data. Qualitative and quantitative results show the acceleration potential for both numerical and in-vivo data for acceleration factors R=1.89, 3.43, and 5.84.

Introduction

The imaging of quantitative MR parameters enables an objective comparison of the investigated tissues on the basis of physical properties and is therefore considered a valuable component of biomarker imaging. Recent work shows that long scan times for qMRI can be reduced to a fraction through data subsampling and model-based reconstructions1. However, reconstruction with nonlinear models can be time-consuming and thus makes them an ideal candidate for deep learning methods. To the best of our knowledge, there have only been very few approaches of mapping MR parameters by means of deep learning and only one, where a model-augmented neural network is used to estimate M0 and T2 maps2. In this work, we propose a neural network that estimates M0 and T1 maps from a corresponding set of subsampled variable flip angle (VFA) images, following the proposed physical model consistency term2 that incorporates the signal model into the objective function. We show the acceleration potential on numerical brain phantoms and on retrospectively subsampled in-vivo measurements.

Methods

Figure 1 illustrates the implemented U-Net architecture3 which follows a basic encoder-decoder structure. The combination of feature information from the contracting path and spatial information from the expansive path justifies the use of a U-Net for this specific problem, where several input images are mapped to only two parameters maps.
As an input the network receives a set of 10 complex images, separated into real and imaginary channels, from a spoiled gradient echo sequence with flip angles {1°:2°:19°} to cover a wide range of possible T1 values. It’s target is to predict the pseudo proton density M0 and the longitudinal relaxation time T1 by optimizing the network parameters according to:
$$ \theta^* = \arg\min_{\theta} \lambda_{UNet} || UNet(S_{\alpha,u} | \theta) - (M0,T1) ||_2^2 + \lambda_{data} \sum_{\alpha} || F^{-1} P F S_{\alpha} (UNet (S_{\alpha,u} | \theta ))-S_{\alpha,u} ||_2^2$$
The operators $$$F^{-1}$$$ and $$$F$$$ denote the (inverse) Fourier transform, $$$P$$$ is the sampling pattern. The left part is a classic loss term for the forward pass of the neural network whereas the right term ensures signal model fidelity according to the signal equation of a spoiled gradient echo sequence4:
$$S_{\alpha}(x) = M0 \frac{\sin(\alpha(x))(1-e^{-TR/T1(x)}) }{1-\cos(\alpha(x))e^{-TR/T1(x)}}$$
The network was trained separately for three different scenarios. A retrospective 1D regular undersampling pattern was equally used for all flip angles with 5% fully-sampled k-space center and additionally sampling every 2nd, 4th, or 8th line, yielding effective acceleration rates of R = 1.89, 3.43, and 5.84.
For training, a generic brain phantom was used, where random Gaussian noise similar to in-vivo noise levels was added. The overall data set consisted of 995 slices which were split into 90% training and 10% test data. The feasibility of this approach was further tested with 420 in-vivo slices from four healthy volunteers by means of transfer-learning which has been shown to be a promising tool for accelerated MRI5. Reference parameter maps were generated using iterative model-based reconstruction without spatial regularization1 and flip angle correction within the signal model was used to compensate for B1+ inhomogeneities6,7.

Results

Figure 2 shows exemplary predicted M0 and T1 maps for the subsampled phantom data including an error map denoting the pixel-wise relative error between prediction and reference parameter maps. Table 1 contains the respective normalized root mean squared error (nRMSE) and structural similarity index8 (SSIM) for the estimated parameter maps for three different acceleration factors. Figure 3 depicts exemplary parameter maps for subsampled in-vivo data with quantitative results in Table 2.

Discussion

The proposed approach from Liu et al.2 is adapted to learn the more complex relation between signal intensity and T1 maps with the signal model term ensuring physical validity. The trained network generates M0 and T1 parameter maps in a time span of only ms which accounts for the huge advantage of using deep learning in image reconstruction provided good training data is available. M0 and T1 maps exhibit excellent visual and quantitative accuracy compared to the fully-sampled reference, also supported by the high similarity of the quantitative metrics given in Table 1. Subsampling with R=5.84 yields predictions with increased pixel-wise outliers. Differences in the estimations for T1 and M0 might be due to the higher complexity of the underlying relation between T1 and the signal intensity and less available information in the training data from a real-valued T1 whereas M0 provides additional phase information. Transferring the learned weights to the in-vivo data yields good results considering the limited amount of training data and even allows to incorporate B1+ maps into the signal model. Minor backfolding artifacts are visible in the difference maps which appear less prominent due to noise for increasing acceleration.

Conclusion

We show the feasibility to estimate M0 and T1 maps from a series of highly undersampled VFA images by means of a neural network. Using transfer-learning the network was fine-tuned to work on in-vivo data, demonstrating its clinical potential for future applications even in the case of limited training data. Further research needs to be done on the amount of retrained layers to completely remove backfolding for in-vivo data. Future work could include an extension to multi-coil data and additionally improving B1+ inhomogeneity maps in a separate output channel.

Acknowledgements

We gratefully acknowledge the support of NVIDIA Corporation with the donation of the Titan Xp GPU used for this research.

References

1. Maier O, Schoormans J, Schloegl M, et al. Rapid T1 quantification from high resolution 3D data with model‐based reconstruction. Magn Reson Med. 2019; 81.10.1002/mrm.27502

2. Lui F, Li F, Kijowski R. Model-Augmented Neural neTwork with Incoherent k-space Sampling for efficient MR T2 mapping. Magn Reson Med. 2019; 82:174– 188.

3. Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. In: Navab N, Hornegger J, Wells WM, Frangi AF, editors. Medical Image Computing and Computer‐Assisted Intervention—MICCAI 2015: 18th International Conference, Munich, Germany, October 5‐9, 2015, Proceedings, Part III. Cham, Switzerland: Springer International Publishing; 2015:234–241.

4. Ernst R, Bodenhausen G, Wokaun A. Principles of nuclear magnetic resonance in one and two dimensions. Oxford: ClarendonPress; 1987.

5. Dar S U H, Çukur T. A Transfer-Learning Approach for Accelerated MRI using Deep Neural Networks. CoRR. 2017; arXiv: 1710.02615.

6. Sacolick LI, Wiesinger F, Hancu I, Vogel MW. B1 mapping by Bloch-Siegert shift. Magn Reson Med. 2010;63:1315–1322.

7. Lesch A, Schloegl M, Holler M, et al. Ultrafast 3D Bloch-Siegert B+1-mapping using variational modeling. Magnetic Resonance in Medicine. 2019; 81: 881– 892

8. Wang Z, Bovik A, Sheikh H, Simoncelli E. Image Quality Assessment: From Error Visibility to Structural Similarity. Image Processing, IEEE Transactions. 2004; on. 13. 600 - 612. doi: 10.1109/TIP.2003.819861

Figures

Figure 1: Implemented U-Net architecture. The network receives the complex, undersampled variable flip angle (VFA) images as input and produces two parameter maps M0 and T1 as an output.

Figure 2: The top row shows the undersampled brain phantom input series for R=5.84. Bottom left and right denote reference and estimated T1 and M0 maps for different acceleration factors 1.89, 3.43, and 5.84. Pixel-wise relative errors are shown below.

Figure 3: The top row shows the undersampled in-vivo input series for R=5.84. Bottom left and right depict corresponding reference and estimated T1 and M0 maps for retrospectively undersampled data of a healthy volunteer. The results are displayed for acceleration factors 1.89, 3.43, and 5.84. Pixel-wise relative errors are shown below. These parameter maps were acquired by fine-tuning the pre-trained network with 420 in-vivo slices which were split into 378 training and 42 test slices.

Table 1: nRMSE and SSIM for estimated M0 and T1 maps for acceleration factors 1.89, 3.43, 5.84.

Table 2: nRMSE and SSIM for estimated M0 and T1 maps for fine-tuning in-vivo data for acceleration factors 1.89, 3.43, 5.84.

Proc. Intl. Soc. Mag. Reson. Med. 28 (2020)
3614