Quantitative mapping of MR parameters has shown great potential for tissue characterization but long acquisition times required by conventional techniques limit their widespread adoption in the clinic. Recently, model-based compressive sensing (CS) reconstructions that produce accurate parameter maps from a limited amount of data have been proposed but these techniques require long reconstruction times making them impractical for routine clinical use. In this work, we propose a multi-scale deep ResNet for MR parameter mapping. Experimental results illustrate that the proposed method achieves reconstruction quality comparable to model-based CS approaches with orders of magnitude faster reconstruction times.
The proposed method was applied to T2 mapping with Radial Turbo-Spin-Echo (RADTSE)11 data acquisition. Fig.1(a) illustrates the supervised training procedure where RADTSE data is used to create multi-contrast images for both input and target. The input data was obtained by applying NUFFT12 to undersampled k-space at each echo time (TE) resulting in images with severe undersampling artifacts. The target data was obtained using model-based CS2,5. Note that if fully-sampled TE images can be acquired, they can be used as targets. However, such acquisitions are impractical in most parameter mapping applications. Unlike earlier deep neural network based image reconstruction methods which train on full images, the proposed network is trained on image patches. Patch-wise processing offers several advantages including requiring fewer training datasets, accelerated training, and low GPU memory requirement. The testing/deployment procedure is illustrated in Fig.1(b). Single-scale ResNet (SS-ResNet)13 (Fig.2(a)) and multi-scale ResNet (MS-ResNet) (Fig.2(b)) architectures were considered to test the hypothesis that a larger receptive field would be more capable of reducing global undersampling artifacts. All convolutional layers consisted of 3x3 kernels with 128 filters. The structure of the residual blocks is shown in Fig.2(b). Note that a residual scaling step with a constant scaling factor of 0.1 is used inside each residual block. In MS-ResNet, the down-sampling block uses convolution with stride=2 and the up-sampling block used the efficient sub-pixel convolution14. The networks were initialized randomly and trained using ADAM optimization with a learning rate of 1e-4. All networks were implemented in Torch15.
Data were acquired from 7 consenting subjects on a 1.5T Siemens Aera scanner using RADTSE on the abdomen with echo spacing= 6.0ms, ETL = 32, TR=1.8-2.3s, and 6 lines/TE with 256 readout points/line. Since fully sampled data was not available, model-based CS2,5 was used to generate the target images. 32x32 patches with stride=15 were extracted from the images and background patches were not included in training. Patches were normalized by subtracting the mean value of each echo across all training images and dividing by the corresponding standard deviation. For the abdomen experiment, three networks were trained: two SS-ResNets with 16 and 32 residual blocks, respectively, and an MS-ResNet with 32 residual blocks.
Data were also acquired from 5 consenting subjects on a 3T Siemens Skyra scanner using RADTSE on the brain with echo spacing= 9.0ms, ETL = 16, TR=5s. For each subject, a reference data (256 lines/TE) acquisition was followed by a separate undersampled data (16 lines/TE) acquisition. The reference and undersampled data were reconstructed using NUFFT. Patches were extracted and preprocessed in the same manner as the abdominal experiment. $$$l_2$$$ loss together with a total variation regularizer with weight=0.001 was used to train an MS-ResNet with 32 residual blocks. In both the abdomen and brain experiments, one subject was randomly selected for testing and the remaining subjects were used for training. T2 maps were calculated using the resulting TE images.
1. Doneva, M., Börnert, P., Eggers, H., Stehning, C., Sénégas, J., & Mertins, A. (2010). Compressed sensing reconstruction for magnetic resonance parameter mapping. Magnetic Resonance in Medicine, 64(4), 1114-1120.
2. Huang, C., Graff, C. G., Clarkson, E. W., Bilgin, A., & Altbach, M. I. (2012). T2 mapping from highly undersampled data by reconstruction of principal component coefficient maps using compressed sensing. Magnetic resonance in medicine, 67(5), 1355-1366.
3. Zhao, B., Lu, W., Hitchens, T. K., Lam, F., Ho, C., & Liang, Z. P. (2015). Accelerated MR parameter mapping with low‐rank and sparsity constraints. Magnetic resonance in medicine, 74(2), 489-498.
4. Velikina, J. V., & Samsonov, A. A. (2015). Reconstruction of dynamic image series from undersampled MRI data using data‐driven model consistency condition (MOCCO). Magnetic resonance in medicine, 74(5), 1279-1290.
5. Tamir, J. I., Uecker, M., Chen, W., Lai, P., Alley, M. T., Vasanawala, S. S., & Lustig, M. (2017). T2 shuffling: Sharp, multicontrast, volumetric fast spin‐echo imaging. Magnetic resonance in medicine, 77(1), 180-195.
6. Schlemper, J., Caballero, J., Hajnal, J. V., Price, A., & Rueckert, D. (2017, June). A Deep Cascade of Convolutional Neural Networks for MR Image Reconstruction. In International Conference on Information Processing in Medical Imaging (pp. 647-658). Springer, Cham.
7. Knoll F., Hammernik K., Garwood E., Hirschmann A., Rybak L., Bruno M., Block T., Babb J., Pock T., Sodickson D. K., & Recht M. P. Accelerated knee imaging using a deep learning based reconstruction, Proc. of ISMRM, 2017.
8. Han Y. S., Lee D., Yoo J, & Ye J. C. Accelerated Projection Reconstruction MR imaging using Deep Residual Learning, Proc. of ISMRM, 2017.
9. Hammernik K., Knoll F., Sodickson D. K., & Pock T. L2 or not L2: Impact of Loss Function Design for Deep Learning MRI Reconstruction, Proc. of ISMRM, 2017.
10. Jin, K. H., McCann, M. T., Froustey, E., & Unser, M. (2017). Deep convolutional neural network for inverse problems in imaging. IEEE Transactions on Image Processing, 26(9), 4509-4522.
11. Altbach, M. I., Outwater, E. K., Trouard, T. P., Krupinski, E. A., Theilmann, R. J., Stopeck, A. T., ... & Gmitro, A. F. (2002). Radial fast spin‐echo method for T2‐weighted imaging and T2 mapping of the liver. Journal of Magnetic Resonance Imaging, 16(2), 179-189.
12. web.eecs.umich.edu/~fessler/irt
13. He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770-778).
14. Lim, B., Son, S., Kim, H., Nah, S., & Lee, K. M. (2017, July). Enhanced deep residual networks for single image super-resolution. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops.
15. http://torch.ch/