3637

Image Reconstruction Using Generative Adversarial Networks with MR-Specific Feature Map
Ruiying Liu1, Hongyu Li1, Dong Liang2, Xiaojuan Li3, Chaoyi Zhang1, Peizhou Zhou1, Leslie Ying1, and Xiaoliang Zhang4
1Department of Biomedical Engineering, Department of Electrical Engineering, The State University of New York at Buffalo, Buffalo, NY, United States, 2Paul C. Lauterbur Research Center for Biomedical Imaging, SIAT, CAS,, Shenzhen, China, 3Program of Advanced Musculoskeletal Imaging (PAMI), Cleveland Clinic, Cleveland, OH, United States, 4Department of Biomedical Engineering, The State University of New York at Buffalo, Buffalo, NY, United States

Synopsis

Deep learning methods have demonstrated great potential in image reconstruction due to its ability to learn the non-linearity relationship between the undersampled k-space data and the corresponding desired image. Among these methods, Generative Adversarial Networks (GANs) is known to reconstruct images that are sharper and more realistic-looking. In this abstract, we study whether an MR-specific feature map that is trained on a large number of MRI images and used in the loss function can improve the GAN-based reconstruction. We demonstrate that the MR-specific feature map is superior to the pre-trained feature map typically used for GAN-based reconstruction.

Introduction

Deep learning has recently been studied for reconstructing MR images from undersampled k-space data1-5. Most studies have found that high-frequency details are lost in deep learning-based reconstruction, and images are perceptually unsatisfactory with overly smooth textures. Generative Adversarial Network (GAN) has gained popularity due to its ability to infer photo-realistic natural images6. Although many works have demonstrated superior reconstruction performance using GAN-based compressed sensing (CS)7-9, the reconstruction time is still too long. The end-to-end reconstruction using deep learning avoids the conventional CS iterations and has the benefit of ultrafast online reconstruction. For GAN-based end-to-end reconstruction, loss function for training is very important. Among the existing loss functions, the hybrid function9 considering both pixel-wise mean square error (MSE) and perceptual loss (VGG networks10,11) has shown to achieve better reconstruction details. However, the perceptual loss uses a high-level feature map pre-trained on ImageNet10, which contains natural images only. Although such a perceptual loss function improves image sharpness, it might hallucinate and introduce natural-looking artifacts when applied to MR reconstruction because the feature map is extracted from natural images only. Here we investigate whether the MR-specific feature maps can improve the GAN-based MR reconstruction in replace of a pre-trained feature map from ImageNet. MR knee images are used to obtain the MR-specific feature map trained on a VGG network and to test the reconstruction performance.

Methods

In GAN, there are two networks, a generator G and a discriminator D. The generator can generate high perceptual quality images according to the discriminator, which is a very good classifier to separate realistic and generated images. The loss function used for the generator is a combination of a content and an adversarial term: $$$loss_G=loss_{cont}+\lambda loss_{adv}(1)$$$, where λ=0.001. For the content term, we use a hybrid loss function: $$$loss_{cont}^{hybrid}=\alpha loss_{cont}^{MSE}+\beta loss_{cont}^{Percept}(2)$$$, where $$$\alpha+\beta \approx 1$$$ , which has shown to be superior to MSE or perceptual loss alone9. The first term is the traditional loss function using mean squared error (MSE): $$$loss_{cont}^{MSE}=\frac{1}{n}\sum_{i=1}^{n}\left \| G(x_i;\theta) -y_i \right \|^{2}(3)$$$, where n is the number of training samples, $$$G(x_i;\theta)$$$ is the reconstructed image generated by generator, $$$y_i$$$ is the ground truth image, and $$$x_i$$$ is the corresponding undersampled k-space data. The second term is the feature-based loss function: $$$loss_{cont}^{Percept}=\frac{1}{n}\sum_{i=1}^{n}\left \| \varphi _{i}G(x_i;\theta) -\varphi _{i}y_i\right \|^{2}(4)$$$, which defines the perceptual similarity using the high-level feature maps $$$\varphi _{i}$$$ from the i = 1,…, n training samples. We compare two different feature maps: one is the widely used feature map pre-trained on ImageNet, the other is our proposed MR-specific feature map. The term $$$loss_{adv}$$$ is defined based on the probabilities of the discriminator D overall training datasets as $$$loss_{adv}=\sum_{i=1}^{n}log\left ( 1-D\left ( G(x_i;\theta )) \right ) \right (5)$$$ , where $$$D\left ( G(x_i;\theta ) \right )$$$ is the probability that the reconstructed image is a ground truth image.6 The architecture of GAN is shown in Figure 1. For all existing GAN with pre-trained feature map, only the parameters in sub-networks G and D are learned during training, while keep the VGG sub-network fixed. In contrast, our proposed method learns the parameters in all three networks during training. Therefore, the feature maps generated by the online learned VGG sub-network include unique features specific to MR images.
A total of 100 subjects were randomly selected from the Osteoarthritis Initiative (OAI) (SAG IW TSE with fat suppression, TR/TE=3200/30ms, resolution 0.5mm × 0.36mm × 3mm, a mixture of radiographic KL grades). 80% (2000 images) were used for training and obtain the MR-specific feature maps, 20% (500 images) were for testing to reconstruct images. 2D Poisson random sampling pattern was used with an acceleration factor of 6. For a fair comparison, we used the same epochs for training which takes around 1 day. The hardware specification is CPU i7-8700K (6 cores 12 threads @4.7GHz); Memory 64 GB; GPU 2x NVIDIA GTX 1080Ti.

Results and Discussion

Figure 2 compares the GAN-based reconstructions using two different feature maps. It can be seen that the reconstruction with pre-trained feature maps suffers from loss of sharpness, while the reconstruction using our proposed MR-specific feature maps has sharp edges. The improvement can be seen more clearly in the error images in error images. The corresponding PSNRs and NMSEs shown on the bottom left of each image also indicate the MR-specific feature maps are better. Figures 3 shows the mean and standard deviation of the NMSEs and PSNRs of the two different feature maps, which are based on a total of 500 testing images. The top of bar is the mean value and error bar indicates the standard deviation. All these quantitative metrics indicate that it is worthwhile to retrain the feature map using MR images for MR reconstruction. It is also worth noting that the result depends to some extent on the training data. A larger-scale statistical analysis will be performed in future work.

Conclusion

In this abstract, we investigate the use of an MR-specific feature map for GAN-based image reconstruction. Experimental results show that the proposed feature map is able to reconstruct images superior to those reconstructed with the conventional pre-trained feature maps. Larger data sets will be used for evaluating diagnostic performance and tissue quantification accuracy (cartilage thickness and composition) in future studies.

Acknowledgements

Acknowledgments: This work is supported in part by the National Institute of Health U01EB023829.

References

1. Wang S, et al. Accelerating magnetic resonance imaging via deep learning. ISBI 2016, pp. 514-517.

2. Jin K H. Deep Convolutional Neural Network for Inverse Problems in Imaging. IEEE Transactions on Image Processing, vol. 26, no. 9, pp. 4509-4522, Sept. 2017.

3. Zhu B, et al. Image reconstruction by domain-transform manifold learning. Nature 555, 2018.

4. Lee D, et al. Deep residual learning for compressed sensing MRI. ISBI 2017 pp. 15-18.

5. Schlemper J, et al. A Deep Cascade of Convolutional Neural Networks for Dynamic MR Image Reconstruction. arXiv:1704.02422, 2017.

6. Ledig C. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2017.

7. Mardani M, et al. Deep Generative Adversarial Networks for Compressed Sensing Automates MRI. arXiv: 1706.00051, 2017.

8. Quan T M. Compressed Sensing MRI Reconstruction Using a Generative Adversarial Network With a Cyclic Loss. IEEE Transactions on Medical Imaging, vol. 37, no. 6, pp. 1488-1497, June 2018.

9. Yang G. DAGAN: Deep De-Aliasing Generative Adversarial Networks for Fast Compressed Sensing MRI Reconstruction. IEEE Transactions on Medical Imaging, vol. 37, no. 6, pp. 1310-1321, June 2018.

10. Simonyan K. Very deep convolutional networks for large-scale image recognition, in ICLR, 2015.

11. He K. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. Proceedings of the IEEE International Conference on Computer Vision, 2015, pp. 1026–1034.

Figures

Figure 1: The architecture of GAN network.

Figure 2 Comparison of the reconstructions (top) and the error images (bottom) using two different feature maps in GAN-based MR reconstruction.

Figure 3: Quantitative comparison, in terms of PSNR and NMSE, between GAN-based reconstructions using two different feature maps.

Proc. Intl. Soc. Mag. Reson. Med. 28 (2020)
3637