4993

Assessing the Role of Deep Learning in Joint Motion and Image Estimation
Brian Nghiem1,2, Zhe Wu1, Melissa Haskell3, Lars Kasper1, and Kamil Uludag1,2
1BRAIN-To Lab, University Health Network, Toronto, ON, Canada, 2Medical Biophysics, University of Toronto, Toronto, ON, Canada, 3Electrical Engineering and Computer Science, University of Michigan, Ann Arbour, MI, United States

Synopsis

We investigated the performance of CNN-assisted joint estimation in two cases of severe motion corruption in a 2D slice of T2w FSE MRI. We showed that the inclusion of the CNN can help speed up convergence of the joint estimation algorithm, corroborating previous findings. We also showed one case in which joint estimation failed to converge to the correct image and motion parameters, with and without the CNN. A more exhaustive study is required to confirm whether deep learning can help joint estimation salvage otherwise unsalvageable corrupted data.

Introduction

Subject motion is a common source of image artifacts that can severely decrease image quality and preclude clinical diagnosis in certain patient cohorts. Data-driven retrospective motion correction methods1–3 are promising alternatives to rescanning that can be readily applied to clinical protocols to salvage corrupted images. While these approaches depend on classical optimization algorithms, recent advances leverage deep learning4 to incorporate useful image priors to assist with the high-dimensional, nonlinear optimization problem. Although the inclusion of deep learning has been shown to improve convergence speed4, it remains unclear whether deep learning can enable image recovery from heavily corrupted data that would otherwise be unsalvageable. In this abstract, we assess the performance limit of classical joint motion & image estimation and investigate the role of deep learning in severe motion corruption cases.

Methods

CNN-Assisted Joint Motion and Image Estimation
The motion-corrupted signal encoding process can be modeled as follows5:
$$ s =E_{\theta} m + \eta \tag{1}$$
$$ E_{\theta} = UFC\theta \tag{2}$$
where $$$s$$$ is the MR signal, $$$m$$$ is the magnetization image, $$$\eta$$$ is Gaussian noise, $$$U$$$ is the k-space undersampling matrix, $$$F$$$ is the Fourier transform, $$$C$$$ denotes the coil sensitivity profiles, and $$$\theta$$$ is composed of a rotation and a translation operator.
Given only the motion-corrupted signal, data-driven retrospective motion correction4 carries out motion-compensated image reconstruction (Eq. 3) by jointly estimating $$$m$$$ and $$$\theta$$$. This can be done using a data consistency formulation (Eq. 4) through the implicit encoding of motion afforded by multi-coil acquisition3.
$$ \hat{m} = \arg\min_{m} || E_{\hat {\theta}} m- s ||^2 \tag{3}$$
$$\hat{\theta} = \arg\min_{\theta} || E_{\theta} \hat{m} - s ||^2 \tag{4}$$
Eq. 3–4 are jointly estimated through coordinate gradient descent3,4 using CG-SENSE6 and the BFGS optimization algorithm [7], respectively. The convergence of the joint estimation algorithm can be hindered by the coupling of $$$m$$$ and $$$\theta$$$ through the interdependence of Eq. 3 and 44. A recent variant of the joint estimation framework (NAMER4), improved convergence behaviour by including a convolutional neural network (CNN) that was trained to estimate artifacts within a motion-corrupted input (Fig. 1). The estimated artifacts are subtracted to produce an artifact-reduced image (Eq. 5), which is then used in place of $$$\hat{m}$$$ in Eq. 4. We reimplemented the NAMER algorithm4 in Python using SigPy, CuPy, and TensorFlow.

$$\hat{m}^* = \hat{m} - CNN(\hat{m}) \tag{5}$$

Data
To facilitate reproducibility, simulations were conducted using a 2D slice of a T2w FSE brain MRI (3T Siemens, 0.49 mm isotropic in-plane resolution) and a motion trajectory provided by the authors of Reference 4 (max translation: $$$1.7 mm$$$; max rotation: 2.4 $$$2.4^\circ$$$). In-plane motion corruption was simulated using the signal encoding model (Eq. 1) with a $$$4x$$$- and $$$5x$$$- rescaling of the motion trajectory shown in Fig 2.a). and an R=2 Cartesian k-space undersampling pattern (Fig 2.b).

Assessing the Performance of Joint Estimation With and Without CNN
With the rescaled motion trajectories described above, we first investigated our baseline ability to reconstruct the uncorrupted image (Eq. 3) when provided with the groundtruth motion parameters.
Continuing with the same motion corruption cases, we compared the convergence behaviour of the joint estimation algorithm with and without the CNN. We analyzed the motion estimation optimization trajectory and assessed the impact of including the CNN.

Results

For all motion-corrupted data, the CG-SENSE algorithm was able to accurately recover the uncorrupted image when provided with the groundtruth motion parameters (Fig. 3). After 100 iterations, reconstructions errors of 1.8%, and 2.4% were achieved for the $$$4x$$$- and $$$5x$$$- rescaled motion trajectories. The number of CG-SENSE iterations required to reach convergence increased as motion severity increased.

In Fig. 4, we show the results of the outputs of the joint estimation algorithm (with and without the CNN) for the $$$4x$$$- and $$$5x$$$- rescaled motion trajectory. For the former case, the algorithm converged to an accurate estimate of the uncorrupted image ($$$\epsilon$$$) with and without the CNN (Fig. 4). For the latter case, the algorithm converged to images with significant residual motion artifacts with and without the CNN, resulting in no improvement in computed reconstruction error.

In Fig. 5, we show the corresponding motion estimation convergence curves for the results shown in Fig. 4. For the $$$4x$$$- rescaled case, we see that including the CNN resulted in faster convergence (by a factor of 1.8). For the $$$5x$$$- rescaled case, we see that motion estimation optimization has converged to suboptimal values with and without the CNN.

Discussion and Conclusions

Here, we presented results of joint estimation with and without the CNN under different examples of severe motion corruption. We had first confirmed that, given the groundtruth motion parameters, we could accurately recover the uncorrupted images (Fig. 3). As such, the differing reconstruction qualities demonstrated in Fig. 4 was due varying levels of success in estimating the motion parameters. This was confirmed by Fig. 5, where we additionally demonstrate the CNN’s potential to speed up convergence towards accurate motion parameter estimates, corroborating previous findings4. From our study, it appears that beyond convergence acceleration, the CNN was not able to salvage motion-corrupted images beyond the joint estimation’s principal capability. However, a more exhaustive study is required to corroborate this finding.

Acknowledgements

References

[1] D. Atkinson, D. L. G. Hill, P. N. R. Stoyle, P. E. Summers, and S. F. Keevil, “Automatic correction of motion artifacts in magnetic resonance images using an entropy focus criterion,” IEEE Trans. Med. Imaging, vol. 16, no. 6, pp. 903–910, 1997

[2] A. Loktyushin, H. Nickisch, R. Pohmann, and B. Schölkopf, “Blind retrospective motion correction of MR images,” Magn. Reson. Med., vol. 70, no. 6, pp. 1608–1618, 2013

[3] L. Cordero-Grande, R. P. A. G. Teixeira, E. J. Hughes, J. Hutter, A. N. Price, and J. V. Hajnal, “Sensitivity Encoding for Aligned Multishot Magnetic Resonance Reconstruction,” IEEE Trans. Comput. Imaging, vol. 2, no. 3, pp. 266–280, 2016

[4] M. W. Haskell et al., “Network Accelerated Motion Estimation and Reduction (NAMER): Convolutional neural network guided retrospective motion correction using a separable motion model,” Magn. Reson. Med., vol. 82, no. 4, pp. 1452–1461, 2019

[5] P. G. Batchelor, D. Atkinson, P. Irarrazaval, D. L. G. Hill, J. Hajnal, and D. Larkman, “Matrix description of general motion correction applied to multishot images,” Magn. Reson. Med., vol. 54, no. 5, pp. 1273–1280, 2005

[6K. P. Pruessmann, M. Weiger, M. B. Scheidegger, and P. Boesiger, “SENSE: Sensitivity encoding for fast MRI,” Magn. Reson. Med., vol. 42, no. 5, pp. 952–962, Nov. 1999

[7] J. Nocedal and S. J. Wright, Numerical optimization, 2nd ed. New York: Springer, 2006.

[8] L. Cordero‐Grande, G. Ferrazzi, R. P. A. G. Teixeira, J. O’Muircheartaigh, A. N. Price, and J. V. Hajnal, “Motion-corrected MRI with DISORDER: Distributed and incoherent sample orders for reconstruction deblurring using encoding redundancy,” Magn. Reson. Med., vol. 84, no. 2, pp. 713–726, 2020

Figures

The NAMER algorithm4 is composed of 3 blocks. The first block removes image features that the CNN has identified as motion artifacts (Eq. 5), producing an artifact-reduced image ($$$\hat{m}^*$$$). The second block estimates the motion parameters ($$$\hat{\theta}$$$) that minimize the data consistency error (Eq. 4) between the measured signal and $$$E_{\theta} \hat{m}^*$$$. The third block reconstructs the image with the updated $$$\hat{\theta}$$$ (Eq. 3). The algorithm loops through these 3 blocks until convergence or a pre-set maximum number of iterations.

a) An in-plane motion trajectory from the Alzheimer's Disease Neuroimaging Initiative (ADNI)), which has a maximum translation magnitude of $$$1.7 mm$$$ and maximum rotation magnitude of $$$2.4^\circ$$$. This trajectory was provided by the authors of Reference 4. b) A Cartesian undersampling pattern, with an acceleration factor of R=2. The sampling profile was chosen to ensure each shot has good k-space coverage along phase encode direction.

We simulated motion-corruption for 3 rescalings ($$$a_{1}=1, a_{2}=4, a_{3}=5$$$) of the motion trajectory depicted in Fig 2.a). For each motion-corrupted image, we carried out motion-compensated image reconstruction (Eq. 3) given the corresponding groundtruth motion parameters ($$$\theta$$$). The CG-SENSE algorithm was able to accurately converge to the uncorrupted image for all three cases.

With the $$$4x$$$- and $$$5x$$$- rescaled motion trajectory, we simulated motion and carried out joint estimation with and without the artifact-removal CNN. For the $$$4x$$$-rescaled case, the algorithm was able to converge to an accurate estimate of the uncorrupted image with and without the CNN. For the $$$5x$$$-rescaled case, the algorithm failed to converge to the uncorrupted image with and without the CNN.

We computed the RMSE of the motion estimates across all joint estimation iterations for both the $$$4x$$$- and $$$5x$$$- rescaled motion trajectories. For the former case, we see that the CNN helped the joint estimation algorithm converge faster. For the latter case, we see that the algorithm converges to inaccurate motion estimates with and without the CNN.

Proc. Intl. Soc. Mag. Reson. Med. 30 (2022)
4993
DOI: https://doi.org/10.58530/2022/4993