3603

ODE-based Deep Network for MRI Reconstruction
Ali Pour Yazdanpanah1,2, Onur Afacan1,2, and Simon K. Warfield1,2
1Harvard Medical School, Boston, MA, United States, 2Boston Children's Hospital, Boston, MA, United States

Synopsis

Fast data acquisition in MRI is vastly in demand and scan time depends on the number of acquired k-space samples. The data-driven methods based on deep networks have resulted in promising improvements, compared to the conventional methods. The connection between deep network and Ordinary Differential Equation (ODE) has been studied recently. Here, we propose an ODE-based deep network for MRI reconstruction to enable the rapid acquisition of MR images with improved image quality. Our results with undersampled data demonstrate that our method can deliver higher quality images in comparison to the reconstruction methods based on the UNet and Residual networks.

Introduction

The connection between deep neural network and the Ordinary Differential Equation (ODE) has been studied and discussed in different recent works [1,2,3,4,5,6]. It has been shown that residual networks such as ResNet [7] and recurrent neural network decoders can be modeled as a discretization of a continuous ODE model. An ODE-based model and its relation to the residual network can be shown as follows:
$$\text L_{t_1} =\text L_{t_0} + f(\text L_{t_0},\theta) \ \ \ (ResNet) \ \ \ (1)$$
$$\text L_{t_1} =\text L_{t_0} + \int_{t_0}^{t_1} f(\text L_{t},\theta)\ dt \ \ \ (ODE) \ \ \ (2)$$
where $$$\text L_{t_0}, \text L_{t_1}$$$ are the residual block input and output. $$$f$$$ represents the network-defined nonlinear operator which preserves the dimensionality of $$$\text L_{t_0}$$$ and $$$\theta$$$ represents the network weights. The defined ODE ($$$\frac{d\text L}{dt} = f(\text L_{t},\theta)$$$) is described in terms of its solution in $$$t=t_1$$$. The forward step of ODE Euler discretization is as follows:
$$\text L_{t_{0}+h} =\text L_{t_0} + hf(\text L_{t_0},\theta) \ \ \ (3)$$
It can be observed that the single forward step of Eq.3, can be considered as the equivalent to the formulation of the residual block. Therefore, The ODE discretization model can lead to different ODE-inspired network architectures. In this paper, we present an ODE-based deep network for Magnetic Resonance Imaging (MRI) reconstruction which extends the conventional reconstruction framework to its data-adaptive variant using the ODE-based network and provides an end-to-end reconstruction scheme. We evaluate the reconstruction performance of our method to the reconstruction methods based on the standard UNet network [8] and Residual network.

Methods

The discretized version of MR imaging model given by
$$\text d=\text E \text x + \text n. \ \ \ (4)$$
where $$$\text x$$$ is the samples of unknown MR image, and $$$\text d$$$ is the undersampled k-space data. $$$\text E =\text {FS}$$$ is an encoding matrix, and $$$\text F$$$ is an undersampled Fourier operator. $$$\text S$$$ is a matrix representing the sensitivity map of the coils, and $$$\text n$$$ is noise. Assuming that the interchannel noise covariance has been whitened, the reconstruction relies on the least-square approach:
$$\hat{\text x} =\underset{\text x}{ argmin} \ \|\text d-\text E\text x\|_{2}^{2} \ \ \ (5)$$
The ODE-based reconstruction framework we used for solving the Eq.5 is shown in Figure 1.
For a conventional neural network, we minimize the loss function ($$$l$$$) over a set of training pairs and we search for the weights ($$$\theta$$$) that minimize that loss function:
$$\underset{\theta}{ minimize} \ \frac{1}{M} \sum\limits_{i=1}^M l(L(\theta;x_i,y_i)) + \text R(\theta) \ \ \ (6)$$
where $$$(x_i,y_i)$$$ is the $$$i_{th}$$$ training pairs (input and ground truth). $$$R$$$ is a regularization operator and $$$M$$$ is the number of training pairs. The loss function depends implicitly on $$$\theta$$$. This optimization is usually solved through Stochastic Gradient Descent (SGD) and backpropagation to compute the gradient of $$$L$$$ with respect to $$$\theta$$$. In our ODE-based network, besides the $$$L$$$s, the network weights also change with respect to time as well. In this case, we need to solve the following constrained optimization problem:
$$\underset{p,\theta}{ minimize} \ \frac{1}{M} \sum\limits_{i=1}^M l(L_{t_1};x_i,y_i) + \text R(p,\theta) \ \ \ (7)$$
$$\frac{d\text L}{dt} = f(\text L_{t},\theta_{t}) \ \ \ (8)$$
$$\frac{d\theta}{dt} = w(\theta_{t},p) \ \ \ (9)$$
where $$$\theta_t$$$ is parameterized by the learnable dynamics of
$$\theta_{t_1} =\theta_{t_0} + \int_{t_0}^{t_1} w(\theta_{t},p)\ dt \ \ \ (10)$$
where $$$w$$$ is a nonlinear operator responsible for the network weights dynamics and $$$p$$$ is the parameters for $$$w$$$. We also augment the learnable space and solve the ODE flow as follows so that the learned ODE representation won't preserve the input space topology [5].
$$\frac{d}{dt} \begin{bmatrix} \text L\\ a \end{bmatrix} = f(\begin{bmatrix} \text L_{t}\\ a_{t} \end{bmatrix},\theta_{t}) \ \ \ (11)$$
where $$$a_{0}=0$$$. We use the discretize-then-optimize method [4,6] to calculate the gradients for backpropagating through ODE layers. Figure 2 shows the proposed ODE-based deep network. Five residual blocks have been used in our method (N=5).

Results

In our experiments, we have tested our method with our MPRAGE brain datasets. The data on ten volunteers with a total of 750 brain images used as the training set. Images from fifteen different patients have used as the testing set. The sensitivity maps were computed from a block of size 24x24 using the ESPIRiT [9] method. Reconstruction results with the undersampling factor of 2x2 for different approaches are shown in Figure 3. ResidualNet includes the same number of residual blocks as our proposed method (without ODE layers). Table 1 shows that our method consistently has higher Peak Signal-to-noise Ratios (PSNR) and structural similarities (SSIM) compared to the reconstructions using the other two networks.

Conclusion

In conclusion, an ODE-based deep network for MRI reconstruction is proposed. It enables the rapid acquisition of MR images with improved image quality. The proposed ODE-based network can be easily adopted by unrolled optimization schemes for better MRI reconstruction accuracy.

Acknowledgements

This research was supported in part by NIH grants R01 NS079788, R01 EB019483, R01 DK100404, and by a research grant from the Boston Children's Hospital Translational Research Program and a pilot grant (PP-1905-34002) from National Multiple Sclerosis Society.

References

[1] Haber, E. and Ruthotto, L., “Stable architectures for deep neural networks,” Inverse Problems, 34(1), 014004 (2017).

[2] Ruthotto, L. and Haber, E., “Deep neural networks motivated by partial differential equations,” arXiv preprint arXiv:1804.04272 (2018).

[3] Chen, T. Q., Rubanova, Y., Bettencourt, J., and Duvenaud, D. K., “Neural ordinary differential equations,” in [Advances in neural information processing systems], 6571–6583 (2018).

[4] Gholami, A., Keutzer, K., and Biros, G., “Anode: Unconditionally accurate memory-efficient gradients for neural odes,” arXiv preprint arXiv:1902.10298 (2019).

[5] Dupont, E., Doucet, A., and Teh, Y. W., “Augmented neural odes,” arXiv preprint arXiv:1904.01681 (2019).

[6] Zhang, T., Yao, Z., Gholami, A., Keutzer, K., Gonzalez, J., Biros, G., and Mahoney, M., “Anodev2: A coupled neural ode evolution framework,” arXiv preprint arXiv:1906.04596 (2019).

[7] He, K., Zhang, X., Ren, S., and Sun, J., “Deep residual learning for image recognition,” in [Proceedings of the IEEE conference on computer vision and pattern recognition], 770–778 (2016).

[8] Ronneberger, O., Fischer, P., and Brox, T., “U-net: Convolutional networks for biomedical image segmentation,” in [International Conference on Medical image computing and computer-assisted intervention], 234–241, Springer (2015).

[9] M. Uecker, P. Lai, M. J. Murphy, P. Virtue, M. Elad, J. M.Pauly, S. S. Vasanawala, and M. Lustig, “Espirit an eigenvalue approach to autocalibrating parallel mri: where sense meets grappa,” Magnetic resonance in medicine, 71(3), 990–1001 (2014).

Figures

The Reconstruction Framework.

The proposed ODE-based deep network. Five residual blocks have been used in our method (N=5).

First row (left to right): Reference image using fully sampled data, ResidualNet reconstruction, UNet reconstruction, and our reconstruction all with an undersampling factor of 2x2. The second row includes error maps correspond to each reconstruction results for comparison.

PSNR and SSIM variations on Brain dataset. Our method consistently has higher PSNR and SSIM compared to the reconstructions using the UNet and Residual networks.

Proc. Intl. Soc. Mag. Reson. Med. 28 (2020)
3603