0226

Joint Reconstruction of MR Image and Coil Sensitivity Maps using Deep Model-based Network
Yohan Jun1, Hyungseob Shin1, Taejoon Eo1, and Dosik Hwang1
1Electrical and Electronic Engineering, Yonsei University, Seoul, Korea, Republic of

Synopsis

We propose a Joint Deep Model-based MR Image and Coil Sensitivity Reconstruction Network (Joint-ICNet), which jointly reconstructs an MR image and coil sensitivity maps from undersampled multi-coil k-space data using deep learning networks combined with MR physical models. Joint-ICNet has two blocks, where one is an MR image reconstruction block that reconstructs an MR image from undersampled k-space data and the other is a coil sensitivity reconstruction block that estimates coil sensitivity from undersampled k-space data. The desired MR image and coil sensitivity maps can be obtained by sequentially estimating them with two blocks based on the unrolled network architecture.

Introduction

In recent years, several methods have been proposed that used deep learning algorithms in an MR image reconstruction for fast MRI1-6. Most of the current deep-learning-based MR reconstruction methods used the coil sensitivity maps that were pre-computed by coil sensitivity estimation methods such as ESPIRiT method7. However, with a high reduction factor or with fewer ACS lines of acquired k-space data, estimated coil sensitivity maps may not be accurate and could affect the reconstruction performance of MR images8. In this study, we propose a Joint Deep Model-based MR Image and Coil Sensitivity Reconstruction Network (Joint-ICNet) that jointly reconstructs an MR image and coil sensitivity maps from undersampled multi-coil k-space data using deep learning networks and interleaved MR model-based data consistency schemes.

Methods

Problem Formulation: The purpose of this study is to reconstruct an MR image from undersampled multi-channel k-space data using deep-learning networks. Thus, the objective function can be formulated as the following least squares equation:
$$\min_{\mathbf{x},\mathbf{C}}\left\Vert{\mathbf{Ax-b}}\right\Vert^2_2+\lambda_I\left\Vert{\mathbf{x}-\mathcal{D}_I(\mathbf{x})}\right\Vert^2_2+\lambda_F\left\Vert{\mathbf{x}-\mathcal{F}^{-1}\mathcal{D}_F(\mathbf{f})}\right\Vert^2_2+\lambda_C\left\Vert{\mathbf{C}-\mathcal{D}_C(\mathcal{F}^{-1}\mathbf{b})}\right\Vert^2_2,$$
where $$$\mathbf{x}$$$ denotes the desired MR image, $$$\mathbf{b}$$$ denotes the measured multi-coil k-space data, $$$\mathbf{A}$$$ denotes the forward operator that has a coil sensitivity map $$$\mathbf{C}$$$ , Fourier transform $$$\mathcal{F}$$$, and a k-space sampling matrix $$$\mathbf{M}$$$, $$$\mathcal{D}_I$$$ is the de-aliasing model that reconstructs artifacts-free MR images from $$$\mathbf{x}$$$, $$$\mathcal{D}_F$$$ is the k-space model that interpolates missing k-space data points from $$$\mathbf{f}$$$, $$$\mathbf{f}$$$ represents the k-space of $$$\mathbf{x}$$$ that is Fourier transformed by $$$\mathcal{F}$$$, $$$\mathcal{D}_C$$$ is the coil sensitivity model that estimates coil sensitivity maps from the acquired k-space data $$$\mathbf{b}$$$, $$$\lambda_I$$$, $$$\lambda_F$$$, and $$$\lambda_C$$$ represent the regularization parameters of $$$\mathcal{D}_I$$$, $$$\mathcal{D}_F$$$, and $$$\mathcal{D}_C$$$, respectively. The least squares problem of the equation can be solved sequentially using a gradient descent method with an iterative algorithm in terms of $$$\mathbf{x}$$$ and $$$\mathbf{C}$$$, respectively, as
$$\begin{cases}\mathbf{x}_{k+1}=\mathbf{x}_k-2\mu_k\left[\mathbf{A}^*(\mathbf{A}\mathbf{x}_k-\mathbf{b})+\lambda_k^I(\mathbf{x}_k-\mathcal{D}_I(\mathbf{x}_k))+\lambda_k^F(\mathbf{x}_k-\mathcal{F}^{-1}\mathcal{D}_F(\mathbf{f}))\right],\\\mathbf{C}_{k+1}=\mathbf{C}_k-2\nu_k\left[\mathbf{x}_k^*\mathcal{F}^{-1}\mathbf{M}^\top(\mathbf{A}\mathbf{x}_k-\mathbf{b})+\lambda_k^C(\mathbf{C}_k-\mathcal{D}_C(\mathcal{F}^{-1}\mathbf{b}))\right],\end{cases}$$
where $$$\mu_k$$$ and $$$\nu_k$$$ are the step size of $$$\mathbf{x}_k$$$ and $$$\mathbf{C}_k$$$ at iteration $$$k$$$ $$$(k = 0, … , N_k)$$$, respectively, and $$$N_k$$$ represents the number of iterations. $$$\mathbf{x}_k$$$ and $$$\mathbf{C}_k$$$ are sequentially reconstructed from above equations.

Proposed Model: The detailed architecture of Joint-ICNet is presented in Fig. 1 and Fig. 2. Joint-ICNet consists of two main blocks: 1) an MR image reconstruction block that reconstructs an MR image from undersampled multi-coil k-space data with convolutional neural network (CNN)-based regularizations and a model-based data consistency layer of an MR image, and 2) a coil sensitivity maps reconstruction block that estimates coil sensitivity maps from undersampled k-space data with a CNN-based regularization and a model-based data consistency layer of coil sensitivity maps. The desired MR image and coil sensitivity maps can be obtained by sequentially estimating them with two blocks.

Implementation Details: We used U-net9 based architectures for the CNN-based regularizations $$$\mathcal{D}_I$$$, $$$\mathcal{D}_F$$$, and $$$\mathcal{D}_C$$$, where U-net has four pooling and up-sampling layers. Each convolutional block located in U-net consisted of a 2D convolutional layer followed by a leaky rectified linear unit (leaky ReLU) with 0.2 negative slope coefficient and instance normalization. The trainable parameters of regularization $$$\lambda_k$$$, and step sizes $$$\mu_k$$$ and $$$\nu_k$$$ were initialized as 1. A total of 10 iterations (i.e., $$$N_k=10$$$) was performed in the unrolled network. The Joint-ICNet was trained with a structural similarity (SSIM) loss function10.

Experiments: We used multi-coil k-space data of the fastMRI open dataset11. The dataset has brain MRI scans including T1 weighted, T1 weighted with contrast agent (T1POST), T2 weighted, and FLAIR images taken from various vendors. During the training, 20% of the slices were used for the training and the validation, which were randomly selected from the dataset. For the test, the whole test slices were used. The k-space data were retrospectively undersampled using 1D Cartesian equispaced sampling masks. The reduction factors were R = 4 and 8. Three metrics including a normalized root mean squared error (NRMSE), a peak signal-to-noise ratio (PSNR), and SSIM were used to quantitatively evaluate the reconstructed images.

Results

We compared Joint-ICNet with a conventional PI-based method, ESPIRiT7, and deep-learning-based MR reconstruction methods, which were U-net9, k-space learning2, and DeepCascade5. Fig. 3 presents the fully sampled and reconstructed (a) FLAIR images with the reduction factor R = 4, and (b) T2 images with the reduction factor R = 8, respectively. Compared to other methods, Joint-ICNet shows superior performance in reconstructing images and removing artifacts as shown on the magnified and difference images. The reconstructed results with abnormal cases are shown in Fig. 4, which presents fully sampled and reconstructed (a) T2, (b) T1POST, and (c) FLAIR images, with the reduction factor R = 4. Whereas the conventional PI-based and other deep-learning-based methods could not recover the blurred abnormal tissues, Joint-ICNet recovered those similar to the fully sampled images, as shown on the magnified images. In addition, Joint-ICNet had the lowest NRMSE, and the highest PSNR and SSIM values compared to other reconstruction methods, as presented in Table 1.

Conclusion

We proposed a joint deep model-based MR image and coil sensitivity reconstruction network, called Joint-ICNet, that jointly reconstructs an MR image and coil sensitivity maps from undersampled multi-coil k-space data. Experiments with various MR images and reduction factors demonstrated that our proposed Joint-ICNet showed superior performance compared to conventional PI-based and state-of-the-art deep-learning-based reconstruction methods in reconstructing the MR images.

Acknowledgements

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT (2019R1A2B5B01070488) and Y-BASE R&E Institute a Brain Korea 21, Yonsei University.

References

[1] Jun Y, Eo T, Shin H, Kim T, et al. Parallel imaging in time‐of‐flight magnetic resonance angiography using deep multistream convolutional neural networks. Magn Reson Med. 2019;81(6):3840–3853.

[2] Han Y, Sunwoo L, Leonard YE, et al. k-Space Deep Learning for Accelerated MRI. IEEE Trans. Med. Imag. 2019;39(2):377–386.

[3] Eo T, Jun Y, Kim T, et al. KIKI‐net: cross‐domain convolutional neural networks for reconstructing undersampled magnetic resonance images. Magn Reson Med. 2018;80(5):2188–2201.

[4] Hammernik K, Klatzer T, Kobler E, et al. Learning a variational network for reconstruction of accelerated MRI data. Magn Reson Med. 2018;79(6):3055–3071.

[5] Schlemper J, Caballero J, Hajnal JV, et al. A deep cascade of convolutional neural networks for dynamic MR image reconstruction. IEEE Trans. Med. Imag. 2017;37(2):491–503.

[6] Aggarwal HK, Mani MP, and Jacob M. MoDL: Model-based deep learning architecture for inverse problems. IEEE Trans. Med. Imag. 2018;38(2):394–405.

[7] Uecker M, Lai P, Murphy MJ, et al. ESPIRiT—an eigenvalue approach to autocalibrating parallel MRI: where SENSE meets GRAPPA. Magn Reson Med. 2014;71(3):990–1001.

[8] Ying L and Sheng J. Joint image reconstruction and sensitivity estimation in SENSE (JSENSE). Magn Reson Med. 2007;57(6):1196–1202.

[9] Ronneberger O, Fischer P, and Brox T. U-net: Convolutional networks for biomedical image segmentation. International Conference on Medical image computing and computer-assisted intervention. 2015;234–241.

[10] Wang Z, Bovik AC, Sheikh HR, et al. Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 2004;13(4):600–612.

[11] Zbontar J, Knoll F, Sriram A, et al. fastMRI: An open dataset and benchmarks for accelerated MRI. 2018; arXiv preprint arXiv:1811.08839.

Figures

Fig. 1. Overall framework of our proposed joint deep model-based magnetic resonance image and coil sensitivity reconstruction network (Joint-ICNet).

Fig. 2. Detailed architecture of Joint-ICNet. (a) Unrolled architecture of Joint-ICNet that has two blocks, where one is a magnetic resonance (MR) image reconstruction block and the other is a coil sensitivity maps reconstruction block. (b) MR image reconstruction block has two convolutional neural network (CNN)-based regularizations and a model-based data consistency layer of MR image. (c) Coil sensitivity maps reconstruction block has a CNN-based regularization and a model-based data consistency layer of coil sensitivity map.

Fig. 3. Fully sampled and reconstructed (a) FLAIR images and (b) T2 images using zero-filling, ESPIRiT, U-net, k-space learning, DeepCascade, and Joint-ICNet, with the reduction factors R = 4 and R =8, respectively.

Fig. 4. Fully sampled and reconstructed (a) T2, (b) T1POST, and (c) FLAIR images with abnormal cases using zero-filling, ESPIRiT, U-net, k-space learning, DeepCascade, and Joint-ICNet, with the reduction factor R = 4.

Table 1. Quantitative evaluation of reconstructed magnetic resonance images with two reduction factors.

Proc. Intl. Soc. Mag. Reson. Med. 29 (2021)
0226