0405

Systematic evaluation of the robustness of open-source networks for MRI multicoil reconstruction
Naoto Fujita1, Suguru Yokosawa2, Toru Shirai2, and Yasuhiko Terada1
1Graduate School of Science and Technology, University of Tsukuba, Tsukuba, Japan, 2FUJIFILM Healthcare Corporation, Tokyo, Japan

Synopsis

Keywords: Image Reconstruction, Machine Learning/Artificial Intelligence, Deep Learning Reconstrcution

Deep neural networks (DNNs) for MRI reconstruction often require large datasets for training, but in clinical settings, the domains of datasets are diverse, and the degree to which deep neural networks are the robustness of DNNs to domain differences between training and testing datasets has been an open question. Here, we evaluated the robustness of four open-source multicoil networks to differences in the domain. We found that model-based networks exhibit higher robustness than data-driven networks and that robustness varies across network architectures, even within model-based networks. Our results provide insight into what network architectures are effective for generalization performance.

Introduction

Deep learning (DL)-based MRI reconstruction has the potential for high performance [1]–[3]. Supervised learning models are commonly used [4], and the performance is higher when the dataset domain used for testing (target domain) is closer to the dataset domain used for training (source domain). However, clinical MRI datasets can have various domain differences depending on the facility, scanner, anatomy, sequence, etc. It is practically difficult to prepare training datasets that encompass such diversity, and differences in the distribution of source and target domains are inevitable.
Thus, there is an open question about how robust the proposed reconstruction networks are against these domain differences. To answer this question, reports [5]–[7] have examined the generalization performance of reconstruction networks. In these studies, however, comparisons have been made only among limited categories of models and domain types.
Here we evaluate the generalization of the reconstruction networks across various domains under clinically practical conditions. Specifically, we compare the reconstruction performance between four network models in terms of the number of images, sampling pattern, acceleration factor (AF), noise level, contrast, and anatomical structure. We use publicly-available networks and multicoil datasets to provide research data that is easy to follow up.

Method

Deep neural networks
We used two categories of MRI reconstruction networks: data-driven and model-based networks. The former uses a large amount of data to learn potential mapping relationships from input (zero-filled image or undersampled k-space data) to output artifact-free images. The latter uses a cascade of DNN to unroll the traditional non-DL optimization iteration algorithm. The models used were as follows:

Data-driven model
- U-Net [8]

Model-based model
- Deep Cascade of Convolutional Neural network (DC-CNN) [3]
- Hybrid Cascade [9]
- Variational Network (VarNet) [1]

Traditional iterative method
CG-SENSE [10], [11]

We used mean-squared error (MSE) for training as the loss function and Adam as the optimizer. For U-net, we used the code from the FastMRI Challenge implementation [12] and trained with 200 epochs, and learning-rate was $$$1.0\times10^{-3}$$$. For DC-CNN and Hybrid Cascade, we modified a public code [13] to use the multicoil data and trained with 50 epochs, and the learning-rate was $$$1.0\times10^{-4}$$$. For VarNet, we also used a public code [14] and trained with 50 epochs, and the learning-rate was $$$1.0\times10^{-3}$$$. For CS-SENSE, we used the reconstruction code of Sigpy [15] with the L1-Wavelet.

Dataset and generalization evaluation
We used the public FastMRI dataset [12]. The multicoil k-space data was retrospectively undersampled and used as input. The sensitivity map was estimated from the undersampled data by ESPiRIT [16], and also used as input. We conducted the following single-domain test (dataset domains for training and testing were the same) and cross-domain test (dataset domains were different between training and testing), Experiments 1-4. The detailed conditions are listed in Table 1.

Exp. 1: Number of training image (single-domain test)
The number of training images was changed from 200 to 4660 images, and the learned models were tested for 400 images.
Exp. 2: Sampling pattern and AF (single-domain test)
Four sampling patterns were used. The patterns for training and testing were the same.
Exp. 3: Noise level (cross-domain test)
Gaussian noises were added to train and test datasets with different noise levels.
Exp. 4: Contrast and anatomy (cross-domain test)
We trained with five subsets with different image contrasts and anatomies. We cross-tested against a different subset than the subset used for training.

Results

The results for single-domain tests (Exp. 1 and 2) are shown in Figs. 1 and 2. VarNet performed better than other networks regardless of the number of training images. The model-based networks (VarNet, Hybrid Cascade, and DC-CNN) outperformed the data-driven network (U-Net).
The results for the cross-domain test (Exp. 3) are shown in Fig. 3. All the model-based networks performed better than U-Net. U-Net was most robust to noise-level differences, although it exhibited worse metrics within the range of noise differences used in this study. Among the model-based networks, Hybrid Cascade was the least robust to noise difference, followed by VarNet.
The results for the cross-domain test (Exp. 4) are shown in Fig. 4. VarNet was robust to both contrast and anatomy domain differences. Hybrid Cascade and DC-CNN were robust only to contrast differences. U-Net was not robust in any case.

Discussion

VarNet generally showed the highest robustness among the model-based networks in all tests. Since VarNet uses an iterative gradient descent algorithm, it could have extracted features independent of differences between training and target domains.
Hybrid Cascade showed higher robustness than DC-CNN except for the noise test. The difference between DC-CNN and Hybrid Cascade is that the latter has a cascade CNN in the k-space domain. The result shows that using the k-space CNN effectively improves performance at the cost of potentially reduced robustness to noise levels.
U-Net showed the lowest robustness in all tests. This is probably because U-Net does not have the regularization term and iterative building blocks of the model-based reconstruction models.

Conclusion

Here we validated the generalization performance of multiple open-source networks for MRI reconstruction through single- and cross-domain tests. Our results provide insight into the architectures that work effectively for generalization performance.

Acknowledgements

No acknowledgement found.

References

[1] K. Hammernik et al., “Learning a variational network for reconstruction of accelerated MRI data,” Magn. Reson. Med., vol. 79, no. 6, pp. 3055–3071, 2018.

[2] H. K. Aggarwal, M. P. Mani, and M. Jacob, “MoDL: Model-based deep learning architecture for inverse problems,” IEEE Trans. Med. Imaging, vol. 38, no. 2, pp. 394–405, 2018.

[3] J. Schlemper, J. Caballero, J. V. Hajnal, A. N. Price, and D. Rueckert, “A deep cascade of convolutional neural networks for dynamic MR image reconstruction,” IEEE Trans. Med. Imaging, vol. 37, no. 2, pp. 491–503, 2017.

[4] Y. Chen et al., “AI-based reconstruction for fast MRI—a systematic review and meta-analysis,” Proc. IEEE, vol. 110, no. 2, pp. 224–245, 2022.

[5] K. Hammernik, J. Schlemper, C. Qin, J. Duan, R. M. Summers, and D. Rueckert, “Systematic evaluation of iterative deep neural networks for fast parallel MRI reconstruction with sensitivity-weighted coil combination,” Magn. Reson. Med., vol. 86, no. 4, pp. 1859–1872, 2021.

[6] J. Huang, S. Wang, G. Zhou, W. Hu, and G. Yu, “Evaluation on the generalization of a learned convolutional neural network for MRI reconstruction,” Magn. Reson. Imaging, vol. 87, pp. 38–46, 2022.

[7] “Assessment of the generalization of learned image reconstruction and the potential for transfer learning - Knoll - 2019 - Magnetic Resonance in Medicine - Wiley Online Library.” https://onlinelibrary.wiley.com/doi/full/10.1002/mrm.27355 (accessed Mar. 15, 2022).

[8] O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical image computing and computer-assisted intervention, 2015, pp. 234–241.

[9] R. Souza, R. M. Lebel, and R. Frayne, “A hybrid, dual domain, cascade of convolutional neural networks for magnetic resonance image reconstruction,” in International Conference on Medical Imaging with Deep Learning, 2019, pp. 437–446.

[10] K. P. Pruessmann, M. Weiger, M. B. Scheidegger, and P. Boesiger, “SENSE: sensitivity encoding for fast MRI,” Magn. Reson. Med. Off. J. Int. Soc. Magn. Reson. Med., vol. 42, no. 5, pp. 952–962, 1999.

[11] K. P. Pruessmann, M. Weiger, P. Börnert, and P. Boesiger, “Advances in sensitivity encoding with arbitrary k-space trajectories,” Magn. Reson. Med. Off. J. Int. Soc. Magn. Reson. Med., vol. 46, no. 4, pp. 638–651, 2001.

[12] J. Zbontar et al., “fastMRI: An open dataset and benchmarks for accelerated MRI,” ArXiv Prepr. ArXiv181108839, 2018.

[13] R. Souza, “CD-Deep-Cascade-MR-Reconstruction.” Aug. 11, 2022. Accessed: Nov. 05, 2022. [Online]. Available: https://github.com/rmsouza01/CD-Deep-Cascade-MR-Reconstruction/blob/1ac44ecefbbe229a3359a6101b482af157a80662/README.md

[14] rixez, “Pytorch implementation of Variational Network for Magnetic Resonance Image (MRI) Reconstruction.” Oct. 02, 2022. Accessed: Nov. 05, 2022. [Online]. Available: https://github.com/rixez/pytorch_mri_variationalnetwork

[15] J. B. Martin, F. Ong, J. Ma, J. I. Tamir, M. Lustig, and W. A. Grissom, “SigPy. RF: comprehensive open-source RF pulse design tools for reproducible research,” in Proc. 28th Annual Meeting of ISMRM, 2020.

[16] M. Uecker et al., “ESPIRiT—an eigenvalue approach to autocalibrating parallel MRI: where SENSE meets GRAPPA,” Magn. Reson. Med., vol. 71, no. 3, pp. 990–1001, 2014.

Figures

Table 1: Training and Testing settings. In this study, networks were trained under 16 different conditions. In the single-domain experiment, training and testing were made under the same conditions. In the cross-domain experiment, training and testing were made under different conditions. Asterisk indicates the same conditions were used in testing as used in training.

Fig. 1 Results for Exp. 1. (A) U-Net, DC-CNN, Hybrid Cascade, and VarNet reconstructed images for the number of training images (N) = 200 and 4660. Error images relative to the fully sampled reference image are displayed with x10 scaling. (B) SSIM vs N.

Fig. 2 Results for Exp. 2. (A) U-Net, DC-CNN, Hybrid Cascade, and VarNet reconstructed image. Error images relative to the fully sampled reference image are displayed with x10 scaling. (B) SSIM.

Fig. 3 Results for Exp. 3. (A) U-Net, DC-CNN, Hybrid Cascade, and VarNet reconstructed images for . Error images relative to the fully sampled reference image are displayed with x10 scaling. (B) SSIM.

Fig. 4 Results for Exp. 4. (A) U-Net, DC-CNN, Hybrid Cascade, VarNet, and CG-SENSE reconstructed images for FastMRI knee dataset. (B) U-Net, DC-CNN, Hybrid Cascade, VarNet, and CG-SENSE reconstructed images for FastMRI brain dataset. Error images relative to the fully sampled reference image are displayed with x10 scaling. (C) SSIM.

Proc. Intl. Soc. Mag. Reson. Med. 31 (2023)
0405
DOI: https://doi.org/10.58530/2023/0405