Jing Cheng1, Zhuo-Xu Cui1, Qingyong Zhu1, and Dong Liang1
1Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
Synopsis
Keywords: Machine Learning/Artificial Intelligence, Image Reconstruction
Existing
deep learning-based methods for MR reconstruction mainly use MSE as loss
function to train the network under the assumption that MR images follow the sub-Gaussian distribution, without considering the real distribution of the
images. In this work, we propose a new DL-based method that models the image
distribution with equilibrium Langevin dynamic to converge the distribution,
and trains the network with Wasserstein distance to
approach the real distribution. Experimental results on highly undersampled MR
data demonstrate the superior performance of the proposed method.
Introduction
Deep learning (DL) has been an import tool
in image reconstruction and shown great potential in significantly speeding up
MR imaging1,2. Most DL-based methods for MR reconstruction use the mean square
error (MSE) of the reconstructed image and the ground truth to train the
network model, which assumes that the MR images follow the sub-Gaussian
distribution. However, in practice, the distribution of the MR images is more
complicated than Gaussian distribution. In this work, we propose a novel
reconstruction approach that uses the Langevin dynamics to model the distribution
of MR images. We adopt the deep equilibrium model to converge the distribution,
and use the Wasserstein distance to measure the
distance between the converged distribution and the real distribution.Theory
Langevin dynamics can be used to generate
samples from the posterior3, which can be formulated as follows in MR imaging:$$x_{t+1}\leftarrow x_{t}+\eta _{t}\bigtriangledown_{x_{t}}log\mu (x_{t}|y)+\sqrt{2\eta _{t}}\zeta _{t}, \zeta _{t}\sim N(0,1) (1)$$ where $$$x\in \mathbb{C}^{N}$$$is
the vector of pixels we wish to reconstruct from the k-space data $$$y\in \mathbb{C}^{M}$$$, $$$\mu (x|y)$$$ is the posterior probability, $$$\eta$$$ is the step size, $$$t$$$ is the time point corresponding to iteration
number.
The core problem in Eq.(1) is to compute the posterior
probability $$$log\mu (x|y)$$$. Assuming the
imaging noise is the i.i.d. Gaussian noise, $$$log\mu (y|x)=-\left\| Ax-y\right\|_{2}^{2}$$$, according to Bayes’ rule, $$$log\mu (x|y)=-\left\| Ax-y\right\|_{2}^{2}+R(x)$$$, where $$$R(x)$$$ represents the imaging prior. Therefore, Eq.(1)
follows$$x_{t+1}= x_{t}+\eta _{t}\bigtriangledown (R(x_{t})-\left\| Ax-y\right\|_{2}^{2})+\sqrt{2\eta _{t}}\zeta _{t}=x_{t}-\eta _{t}A^H (Ax-y)+\eta _{t}\bigtriangledown R(x_{t})+\sqrt{2\eta _{t}}\zeta _{t} (2)$$ Eq.(2) is the form of noisy gradient
descent, similar to the gradient descent algorithm. Recalling to the unrolling
methods, deep network is used to approximate the unknown functions in the
imaging model with end-to-end training. Thus, we replace the function $$$\bigtriangledown R(x_{t})$$$ with a parameterized operator $$$g(x)$$$.
To make
the image distribution converge, we adopt the deep equilibrium model (DEQ)4. In DEQ,
the mapping $$$f_{\theta }$$$, corresponding
to the iteration that $$$x_{t+1}=f_{\theta }(x_{t};y)$$$, has
the same parameters in each iteration. The
limit of $$$x_{T}$$$ as $$$T\to \infty $$$, provided
it exists, is a fixed point of the operator $$$f_{\theta }(\cdot ;y)$$$. The
fixed-point $$$x_{T}$$$ is a
good estimate of the image given its measurement $$$y$$$.
With DEQ and end-to-end Langevin dynamics, the
reconstructed image distribution converges. We also use the network
architecture of WGAN with gradient penalty5 to make the converged distribution
approach to the real distribution.Method
T2-weighted MR data from MoDL was used to
evaluate the feasibility of the
proposed method. The raw data were acquired using a 3D T2 CUBE sequence with
12-channel head coil. 360 slices from training subjects were used to train the
model and 164 slices from the test subject for testing. Variable density
psedo-random sampling mask was used to demonstrate the
performance of reconstruction methods.Results
We compared our proposed approach with
different DL-based MR reconstruction methods, including cycleGAN6, DE-GRAD7,
HDSLR8 and MoDL9.
The qualitative comparisons with an acceleration
factor of 12 are shown in Fig 1. The reconstructed images, as well as the
corresponding zoom-in images were provided. The data distribution modeling
methods (proposed and cycleGAN) achieve better performance than the
conventional data sample modeling methods (DE-GRAD, HDSLR and MoDL) due to the
advantage of data distribution modeling. The proposed method can faithfully
reconstruct the images with clearer anatomical details indicated by the zoom-in
images. Quantitative results of different methods with R=12 are presented in
Table I. results reported are on the entire test dataset, thus the values in
the table are the mean values. Figure 2 shows the robustness of the proposed
method when changing the data settings. Specifically, we directly changed the
input of the trained model to random noise, and added 5dB noise to the
measurement. Figure 2(a) illustrates the reconstructions with noise input, and
Fig 2(b) is the reconstructions with noisy measurement. Unrolling-based methods
(HDSLR and MoDL) failed to reconstruct images, and the proposed method achieves
good performance in detail preservation and artifact removal, showing better
robustness. Conclusion
In this work, we have proposed a novel
DL-based method to approximate the real distribution of MR images. Experimental
results show the superior performance of the proposed approach.Acknowledgements
This work was supported in
part by the National Key Research and Development Program of China under Grant 2020YFA0712200; in part by the National
Natural Science Foundation of China under Grants 12026603, U1805261 and 62106252; in part by the Key Laboratory for Magnetic Resonance and Multimodality Imaging of
Guangdong Province under Grant 2020B1212060051.References
[1] Wang
G, Ye J, Mueller K, et al. Image Reconstruction Is a New Frontier of Machine
Learning. IEEE Transactions on Medical Imaging 2018, 37(6):1289-1296.
[2] Liang
D, Cheng J, Ke Z, et al. Deep Magnetic Resonance Image Reconstruction: Inverse
Problems Meet Neural Networks. IEEE Signal Processing Magazine 2020, 37(1):141-151.
[3]
Bakry D andÉmery M. Diffusions
hypercontractives. In Seminaire de probabilités XIX 1983/84, pages 177–206.
Springer, 1985.
[4]
Bai S, Kolter JZ and Koltun V. Deep
equilibrium models. Proc. Adv. Neural Inf. Process. Syst., 690-701, 2019.
[5]
Gulrajani I, Ahmed F, Arjovsky M,
Dumoulin V, and Courville A. Improved training of wasserstein GANs. In
Proceedings of the 31st International Conference on Neural Information
Processing Systems 2017, 5769–5779.
[6]
Oh G, Sim B, Chung H,
et al. Unpaired Deep Learning for Accelerated MRI Using
Optimal Transport Driven CycleGAN. IEEE Transactions on Computational Imaging 2020,
6:1285-1296.
[7]
Gilton D, Ongie G and Willett R.
Deep Equilibrium Architectures for Inverse Problems in Imaging. IEEE
Transactions on Computational Imaging 2021, 7:1123-1133.
[8]
Pramanik A, Aggarwal HK, Jacob
M. Deep Generalization of Structured Low-Rank Algorithms (Deep-SLR). IEEE Transactions
on Medical Imaging 2020, 39(12):4186-4197.
[9]
Aggarwal HK, Mani MP, Jacob M.
MoDL: Model-Based Deep Learning Architecture for Inverse Problems. IEEE Transactions
on Medical Imaging 2019, 38(2):394-405.