3172

Using a deep residual learning framework to enhance APT-weighted (APTw) MR images of brain tumors acquired by SENSE with compressed sensing
Jingpu Wu1,2, Yiqing Shen1,3, Pengfei Guo1,3, Qianqi Huang3, Babak Moghadas1, Hye-Young Heo1, Jinyuan Zhou1, and Shanshan Jiang1
1Department of Radiology, School of Medicine, Johns Hopkins University, Baltimore, MD, United States, 2Department of Applied Mathematics and Statistics, Whiting School of Engineering, Johns Hopkins University, Baltimore, MD, United States, 3Department of Computer Science, Whiting School of Engineering, Johns Hopkins University, Baltimore, MD, United States

Synopsis

Keywords: CEST & MT, CEST & MT

Sensitivity encoding (SENSE) is often adopted to accelerate image acquisition for various MRI sequences, including APTw. This is further accelerated with SENSE with compressed sensing (called CS-SENSE), but the image quality degrades to some extent. We collected both SENSE- and CS-SENSE-APTw images and trained a generative model with residual learning to generate SENSE images from CS-SENSE images. The generated results were proved to be highly similar to SENSE-APTw images and less noisy than both SENSE- and CS-SENSE-APTw images. With a larger dataset, we can train more robust models and eventually replace SENSE- with CS-SENSE for a speedup of ~50%.

Introduction

Amide proton transfer-weighted (APTw) imaging, a type of chemical exchange saturation transfer (CEST) MRI,1-4 is a protein-based molecular imaging technique.5 This technique has been successfully applied to brain tumors and many other diseases.6-11 Currently, clinical APTw imaging has relatively longer acquisition time and is highly sensitive to motion. Sensitivity encoding (SENSE)12 has been widely adopted in APTw imaging to speed up the acquisition, while still preserving satisfactory image quality. Recently, compressed sensing (CS) was further introduced to achieve even faster acquisition (called CS-SENSE), but the image quality degraded with a higher acceleration factor (AF). Previous studies have demonstrated that a CS-SENSE AF of 4 is acceptable for APTw MRI.13, 14 In this work, using SENSE-APTw images with an AF of 2 as the ground truth, we trained a neural network, which takes CS-SENSE-APTw images with an AF of 4 as the input, to generate images towards the ground truth. If the generated images have desirable quality, we may replace SENSE- with CS-SENSE and save nearly half of the acquisition time.

Method

MR imaging was performed on a Philips 3T MRI scanner. A recommended 3D APTw imaging sequence11 (saturation power = 2 μT; saturation time = 2 sec; TR = 6.5 sec; FOV = 212×192 mm2; 15 slices; slice thickness = 4 mm; matrix = 120×118, reconstructed to 256×256; SENSE = 2 or CS-SENSE = 4) was used to acquire APTw images. 63 patients with gliomas (male 41, female 22; age 53.5±13.9 years, ranged from 26-80 years) were included in our study and they were randomly divided into the training (50 patients) and testing (13 patients) sets. For each patient, SENSE- and CS-SENSE-APTw images were co-registered and skull-stripped, and the voxel values were normalized to [0, 1]. Each APTw image has 15 slices which were used as 15 separate 2D images.

CS-SENSE uses undersampling of k-space data to accelerate acquisition and the result is different from that acquired by SENSE, as there is an information loss in the undersampling process. Notably, to restore SENSE images using CS-SENSE images, we did not generate SENSE-APTw images themselves, as it was found harder to train with. Instead, we trained a convolutional autoencoder to learn the difference mapping between paired SENSE and CS-SENSE data (i.e., SENSE minus CS-SENSE). Subsequently, in the inference stage, the generated residuals from the autoencoder were added to the input CS-SENSE-APTw images to get the estimation of SENSE-APTw images. The architecture of the convolutional autoencoder was shown in Figure 1. It used 3 convolutional and transposed convolutional layers for under-sampling and up-sampling, which can extract features at both high and low levels while preserving more texture details than models using pooling layers.

We employed mean squared error (MSE), structural similarity index measure (SSIM), and peak signal-to-noise ratio (PSNR) as the metrics to evaluate an image's similarity to corresponding ground truth (SENSE-APTw image). Lower MSE and higher SSIM and PSNR indicate higher similarity. The standard deviation (SD) of image signal intensities was used to estimate the level of noise. Lower SD indicates lower level of image noise.

Results and Discussion

Two representative examples of the acquired SENSE-APTw and CS-SENSE-APTw images, as well as generated APTw images were illustrated in Figure 2. The first case (with recurrent anaplastic oligodendroglioma) proved the model’s capability to generate a high-quality APTw image using the undersampled one, with a visible noise level reduction; the second case (with treatment effect) demonstrated the model’s robustness to the motion artifact. The average metric values in the testing set were depicted in Table 1. Quantitatively, our model can not only accurately generate the high-quality SENSE-APTw images from CS-SENSE-APTw images, but can also reduce the noise level by approximately 19% compared to SENSE- and CS-SENSE APTw images (Table 2).

The size of the dataset is now small and it should further be verified whether the model can generate desirable results for any possible input. However, the preliminary results show that this work is a good start to our ultimate goal of completely replacing SENSE with CS-SENSE for faster APTw acquisition.

Conclusion

A convolutional autoencoder model with residual learning has been built to generate APTw images from CS-SENSE-APTw images using SENSE-APTw images as the ground truth. The generated images were proved to be highly similar to the ground truth and less noisy than both the input and the ground truth on average.

Acknowledgements

The authors thank our clinical collaborators for help with the patient recruitment and MRI technicians for assistance with MRI scanning. This study was supported in part by grants from the NIH.

References

1. Ward KM, Balaban RS. Magn Reson Med. 2000;44:799-802.

2. van Zijl PCM, Yadav NN. Magn Reson Med. 2011;65:927-948.

3. Vinogradov E, Sherry AD, Lenkinski RE. J Magn Reson. 2013;229:155-172.

4. Jones KM, Pollard AC, Pagel MD. J Magn Reson Imaging. 2018;47:11-27.

5. Zhou J, Payen J, Wilson DA, Traystman RJ, van Zijl PCM. Nature Med. 2003;9:1085-1090.

6. Zhou J, Heo H-Y, Knutsson L, van Zijl PCM, Jiang S. J Magn Reson Imaging. 2019;50:347-364.

7. Jiang S, Eberhart CG, Lim M, et al. Clin Cancer Res. 2019;25:552-561.

8. Sotirios B, Demetriou E, Topriceanu CC, Zakrzewska Z. Eur J Radiol. 2020;133:109353.

9. Herz K, Mueller S, Perlman O, et al. Magn Reson Med. 2021;86:1845-1858.

10. Gao T, Zou C, Li Y, Jiang Z, Tang X, Song X. Int J Mol Sci. 2021;22:doi: 10.3390/ijms222111559.

11. Zhou J, Zaiss M, Knutsson L, et al. Magn Reson Med. 2022;88:546-574.

12. Pruessmann KP, Weiger M, Scheidegger MB, Boesiger P. Magn Reson Med. 1999;42:952-962.

13. Heo HY, Xu X, Jiang S, et al. Magn Reson Med. 2019;82:1812-1821.

14. Zhang N, Zhang H, Gao B, et al. Front Neurosci. 2022;16:876587.

Figures

Figure 1. Architecture of the convolutional autoencoder model.

Figure 2. Two examples of the acquired SENSE-APTw (ground truth) and CS-SENSE-APTw (input) images, as well as generated APTw (output) images.

Table 1. Average metric values of CS-SENSE-APTw and generated APTw images for measuring their similarity to the SENSE APTw images (ground truth) on the testing set.

Table 2. Standard deviations (SD) of the signal intensity histograms of the acquired SENSE-APTw (target) and CS-SENSE-APTw (input) images, as well as generated APTw images (255 bins).

Proc. Intl. Soc. Mag. Reson. Med. 31 (2023)
3172
DOI: https://doi.org/10.58530/2023/3172