0142

Accelerate Parallel CEST Imaging with Dynamic Convolutional Recurrent Neural Network
Huajun She1, Quan Chen1, Shuo Li1, Kang Yan1, Xudong Chen1, Xi Chen1, Yuan Feng1, Jochen Keupp2, Robert Lenkinski3,4, Elena Vinogradov3,4, and Yiping P. Du1

1Institute for Medical Imaging Technology, School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, China, 2Philips Research, Hamburg, Germany, 3Radiology, University of Texas Southwestern Medical Center, Dallas, TX, United States, 4Advanced Imaging Research Center, University of Texas Southwestern Medical Center, Dallas, TX, United States

Synopsis

CEST is a new contrast mechanism in MRI. However, a successful application of CEST is hampered by its slow acquisition. This work investigates accelerating parallel CEST imaging using dynamic convolutional recurrent neural networks. This work is the first try to apply recurrent neural networks to accelerate CEST imaging, which jointly learns the spatial and Z-spectral features. The in vivo brain results show that the proposed method demonstrates a much better reconstruction quality of the human brain MTRasym maps than the traditional dynamic compressed sensing method, while the reconstruction time is one hundred times shorter.

Introduction

Chemical exchange saturation transfer (CEST) is a novel contrast mechanism in MRI1. However, a successful translation of CEST into clinic might be hampered by its time-consuming acquisition. With the successful application of compressed sensing (CS) theory in MRI2,3, several CS techniques have been applied to accelerate CEST4-6. However, CS based methods are computational expensive. Deep learning is the new frontier for MRI reconstruction7,8, and deep convolutional network9 had been tested with CEST10. As recent advancement, a convolutional recurrent neural network (CRNN11) was introduced, which learns the spatial-temporal dependencies in dynamic MRI, and has shown high quality reconstruction in single coil heart imaging11. Here, we extended the framework combining multi-coil data into training network to accelerate parallel CEST-MRI. Using retrospectively sub-sampled in-vivo CEST-MRI data, we demonstrate that the reconstruction quality of the proposed learning framework is much better than the traditional dynamic compressed sensing method3, while the reconstruction time is about one hundred times shorter.

Theory and Methods

The parallel CRNN (PCRNN) network learns the correlations in spatial-temporal domain and the correlations among iterations. For the CEST application, the temporal domain is replaced by the Z-spectral domain, and PCRNN jointly learns the spatial and Z-spectral features. As shown in Fig.1, the input of the initial iteration is composed of multi-coil undersampled images, and the target of the output is multi-coil fully sampled images. The bidirectional CRNN layers (red bars) were used to learn the spatial-Z correlations, and CRNN layers (yellow bars) were used to learn the dependencies among iterations. Then it projects the extracted features to image domain with convolution layers (green bars). The residual layer (orange summation) allows faster convergece12. The DC layer (purple bar) constraints the reconstruction image to be consistent to the undersampled k-space data9. The reconstruction from the i-th iteration is taken as the new input to the next iteration, which will improve the quality of the input over iterations. All the experiments were performed on a Philips 3T scanner using a 32-channel head coil. CEST images were acquired with a TSE sequence, TR/TE=4200/6.4 ms, slice thickness=4.4 mm, matrix=240x240, FOV=240x240 mm. The saturation RF consisted of 40 pulses each of 49.5 ms duration with 0.5 ms intervals; 31 saturation offsets were recorded between ±7.8ppm with one additional image acquired without saturation for normalization. Data from ten healthy volunteers were collected for training, and the data from three other volunteers were used for testing. For each volunteer, three saturation power levels were tested: 0.7μT, 1.2μT, and 1.6μT. Data augmentation was used for the training data to avoid overfitting. CEST processing used WASSR13 for B0 inhomogeneity correction. MTRasym­ maps were calculated at 3.5 ppm (Amide Proton Transfer weighted). Software channel compression14 was performed to combine the 32-coil data into 4 virtual coils. We implemented the learning algorithm with TensorFlow15. All the computations were performed on a workstation with NVIDIA GTX 1080 Ti GPU. Due to the GPU memory limitation, we resized all of the data to 120x120, and a batch size of 1. Dynamic compressed sensing k-t Sparse SENSE3 was used for performance comparison.

Results and Discussion

Fig.2-4 compare the results reconstructed using fully sampled, k-t Sparse SENSE and the PCRNN from datasets of the three volunteers. In Fig.2-4 (a), the reconstruction MTRasym maps and difference images demonstrate that PCRNN can give much better reconstruction quality than k-t Sparse SENSE, the reconstruction error "e" calculated from the normalized square root mean square error6 (nRMSE) is shown on the difference maps. Fig.2-4 (b) shows the Bland-Altman plot16 and correlation plot reconstructed with PCRNN, k-t Sparse SENSE and the fully sampled case for different saturation power in the region of interest (ROI). It is evident that PCRNN leads to higher correlation (r2) with the fully sampled data than k-t Sparse SENSE. Fig.2-4 (b) also shows that PCRNN gives better ROI-averaged Z-spectrum than k-t Sparse SENSE. Importantly, to get these much better quality reconstruction, PCRNN only need around 1 second, which is about one hundred times shorter than k-t Sparse SENSE. Fig.5 shows the testing error of volunteer #1 over training epochs. We can see the testing error decreased fast at first fifty epochs, but does not change much between 120~150 epochs. We evaluated the performance by Cartesian undersampling2, and the acceleration factor was R=4. Evaluations at higher acceleration factors (R=6,8,etc) are underway.

Conclusion

We proposed and evaluated a parallel CRNN (PCRNN) learning framework for accelerating CEST imaging. In-vivo human brain results demonstrate that at acceleration rate R=4, the reconstruction quality of PCRNN is much better than k-t Sparse SENSE, while the reconstruction time is one hundred times shorter.

Acknowledgements

This work was supported by the NIH grant R21 EB020245 and by the UTSW Radiology Research fund. We thank Dr. Asghar Hajibeigi for the CEST experimental preparation. We thank Yufei Zhang for image processing and the valuable discussion on the algorithms.

References

[1] van Zijl P, et al. MRM 2011;65:927–948.

[2] Lustig M, et al. MRM 2007;58:1182–1195.

[3] Otazo R, et al. MRM 2010; 64:767–76.

[4] Heo HY, et al. MRM 2017;77:779–786.

[5] Zhang Y, et al. MRM 2017;77:2225–2238.

[6] She H, et al. MRM 2018;27400.

[7] Zhu B, et al. Nature 2018;555:487–492.

[8] Hammernik K. MRM 2018;79:3055–3071.

[9] Schlemper J, et al. IPMI 2017;647–658.

[10] She H, et al. ISMRM 2018;5112.

[11] Qin C, et al. arXiv:1712.01751.

[12] He K, et al. ICCCV 2015;1026.

[13] Kim M, et al. MRM 2009;61:1441–1450.

[14] Huang F, et al. MRM 2008;26:133–141.

[15] https://www.tensorflow.org

[16] Martin BJ, et al. Lancet 1986;327:307–310.

Figures

Fig. 1 shows the structure of PCRNN. The input is composed of multi-coil (C=4) undersampled images, the output is multi-coil fully sampled images. The network is composed of bidirectional CRNN layers (red bars) followed by CRNN layers (yellow bars), convolution layers (green bars), residual layer (summation), and data consistency (DC) layer (purple bar). Fast Fourier transform (FFT) is taken on the undersampled image for data consistency. The input of the initial iteration is composed of multi-coil undersampled images, and the target of the output is multi-coil fully sampled images. The reconstruction from the i-th iteration is taken as the new input to the next iteration.

Fig.2(a) MTRasym maps reconstruction of fully sampled, k-t Sparse SENSE, PCRNN and Difference maps of volunteer #1 for three saturation powers. (b) Left and Center: The Bland-Altman plot and correlation plot in the region of interest (ROI) for 3 power levels: 1.6μT (red), 1.2μT (green), and 0.7μT (blue). The horizontal axis in Bland-Altman plot represents the pixel value in the ROI of the MTRasym map, and the vertical axis represents the difference between the reconstructed MTRasym map and the fully sampled MTRasym map. The ROI is shown at of the bottom row. (b) Right: ROI-averaged Z-spectra using k-t Sparse SENSE, PCRNN and fully sampled reconstruction.

Fig.3(a) MTRasym maps reconstruction of fully sampled, k-t Sparse SENSE, PCRNN and Difference maps of volunteer #2 for three saturation powers. (b) Left and Center: The Bland-Altman plot and correlation plot in the region of interest (ROI) for 3 power levels: 1.6μT (red), 1.2μT (green), and 0.7μT (blue). The horizontal axis in Bland-Altman plot represents the pixel value in the ROI of the MTRasym map, and the vertical axis represents the difference between the reconstructed MTRasym map and the fully sampled MTRasym map. The ROI is shown at of the bottom row. (b) Right: ROI-averaged Z-spectra using k-t Sparse SENSE, PCRNN and fully sampled reconstruction.

Fig.4(a) MTRasym maps reconstruction of fully sampled, k-t Sparse SENSE, PCRNN and Difference maps of volunteer #3 for three saturation powers. (b) Left and Center: The Bland-Altman plot and correlation plot in the region of interest (ROI) for 3 power levels: 1.6μT (red), 1.2μT (green), and 0.7μT (blue). The horizontal axis in Bland-Altman plot represents the pixel value in the ROI of the MTRasym map, and the vertical axis represents the difference between the reconstructed MTRasym map and the fully sampled MTRasym map. The ROI is shown at of the bottom row. (b) Right: ROI-averaged Z-spectra using k-t Sparse SENSE, PCRNN and fully sampled reconstruction.

Fig. 5 MTRasym maps of PCRNN (top row) and difference maps (bottom row) of volunteer #1 for different epochs. The testing error decreased fast at first fifty epochs, but does not change much between 120 ~ 150 epochs.

Proc. Intl. Soc. Mag. Reson. Med. 27 (2019)
0142