0074

Network Accelerated Motion Estimation and Reduction (NAMER): Accelerating forward model based retrospective motion correction using a convolutional neural network
Melissa W. Haskell1,2, Stephen F. Cauley1,3, Berkin Bilgic1,3, Julian Hossbach4, Josef Pfeuffer4, Kawin Setsompop1,3,5, and Lawrence L. Wald1,3,5

1A.A. Martinos Center for Biomedical Imaging, Charlestown, MA, United States, 2Program in Biophysics, Harvard University, Cambridge, MA, United States, 3Harvard Medical School, Boston, MA, United States, 4Siemens Healthineers AG, Erlangen, Germany, 5Harvard-MIT Division of Health Sciences and Technology, MIT, Cambridge, MA, United States

Synopsis

Retrospective motion correction techniques have the potential to improve clinical imaging without altering the workflow or acquisition sequence. Yet, they suffer from long reconstruction times and poor conditioning. To address these problems, we developed a Network Accelerated Motion Estimation and Reduction method (NAMER) within a data-consistency based forward model approach to motion parameter estimation. The neural net accelerates convergence up to 15-fold as well as improving final image quality. The ML+MR physics motion correction method combines the speedup provided by fast convolutional neural networks with the robustness of a forward model-based data-consistency reconstruction.

Introduction

Patient motion is one of the most common and costly types of MRI artifacts1. A large variety of MRI motion correction techniques exists2, but many methods have failed to gain traction in the clinic due to workflow complications (tracking devices) or acquisition disruptions (navigators). Fully retrospective motion correction methods that use only raw k-space data avoid these pitfalls3–5. However, they present a poorly conditioned, large, nonconvex model inversion problem which must be solved in a clinically acceptable timeframe.

We address these issues by incorporating a convolutional neural network (CNN) step into a SENSE+motion model-based reconstruction. First, the CNN provides an initial guess of the motion artifacts, thereby improving the conditioning of the nonconvex motion search and accelerating convergence. Secondly, the method uses the CNN within the iterative forward model solve to remove motion artifacts from the current iterate image. This approach is similar to those of variational networks, which use a machine learning step to regularize a highly undersampled reconstruction problem6.

Methods

Figure 1 outlines the NAMER (Network Accelerated Motion Estimation and Reduction) method. Motion corrupted k-space data is first reconstructed assuming no motion has occurred using the standard SENSE forward model reconstruction7. Motion correction is then performed by iterating through the following three steps:

1. CNN image update: Figure 2 shows the motion artifact detecting CNN architecture8. Motion corrupted image patches (size 51x51) are passed into the CNN and the artifacts within the patches are detected. Two channel input data was used to hold the real and imaginary components. The network consists of convolutional, batch normalization, and ReLU activation layers (27 total), with 3x3 kernels. Training was performed using simulated motion corrupted images based on a motion forward model9,10. The artifacts detected by the CNN are subtracted from the original input image (i.e. $$$\boldsymbol{x}_{cnn}=\boldsymbol{x}-\mathrm{CNN}(\boldsymbol{x})$$$).

2. Physics-based motion update: Next, we minimize the data consistency error between the acquired k-space data, $$$\boldsymbol{s}$$$, and $$$\boldsymbol{E}_\boldsymbol{\theta}\boldsymbol{x}_{cnn}$$$ (the encoding model for a given motion trajectory, $$$\boldsymbol{\theta}$$$, applied to the CNN generated image, $$$\boldsymbol{x}_{cnn}$$$), across all of the motion parameters:

$$[\hat{\boldsymbol{\theta}}] = \underset{\boldsymbol{\theta}}{\mathrm{argmin}} ||\boldsymbol{s}-\boldsymbol{E}_{\boldsymbol{\theta}}\boldsymbol{x}_{cnn}||_2$$

3. Physics-based image update: Using the motion from the previous step, we then solve for the image using a SENSE+motion forward model:

$$[\hat{\boldsymbol{x}}] = \underset{\boldsymbol{x}}{\mathrm{argmin}} ||\boldsymbol{s}-\boldsymbol{E}_{\boldsymbol{\theta}}\boldsymbol{x}||_2$$

CNN Training: Our CNN was trained on simulated data, generated from in vivo T2-weighted RARE/TSE/FSE11 images with TR/TE=6.1s/103ms, in-plane FOV=220x220mm2, 4mm slices, 448x448x35mm3 matrix size, 0.5x0.5mm2 in-plane resolution, R=2 undersampling, and TF=11. Ten evenly spaced slices from four healthy subjects, each corrupted by 10 different motion trajectories were used as training data (400 examples total). Motion trajectories were taken from AD patient fMRI runs and augmented using shifting and scaling.

Validation: The CNN was applied to simulated motion corrupted slices from a fifth healthy subject to assess the performance on unseen anatomy. For a test slice, twenty NAMER iterations were performed. To compare to standard iterative methods without the CNN, an alternating minimization4,12 was performed until convergence (about 80 iterations). Lastly, NAMER was applied to an additional two "real" motion cases, where the subjects were asked to shake their heads "no" during the middle of the scan.

Results

Figure 3 shows the performance of the CNN alone to remove image artifacts across the whole brain of an unseen test subject. All slices had reduced image space RMSE, with an average of 3.9%. While the CNN alone removes some motion artifacts, it is incomplete and introduces some blurring.

Figure 4 shows the data-consistency error of the motion search at each iteration of the NAMER and alternating methods. NAMER converges to a high-quality image after five iterations, while the no-CNN method still has significant artifacts. To reach a data-consistency error of 9.8% (close to the SNR limit of the acquisition), NAMER took 6 iterations, whereas the alternating method required 60, leading to a >15x increase in reconstruction time.

Figure 5 shows the image results from two real motion cases. Ringing was reduced for both subjects, and the position coordinates returned from NAMER correspond to the expected motion, based on the instructions given to the subjects.

Discussion and Conclustion

The NAMER method introduces a CNN to improve and accelerate iterative model-based retrospective motion correction. This method provides the confidence of a classical physics-based reconstruction with the computational benefits of a ML network. The CNN used in NAMER generalized to unseen anatomy and successfully removed artifacts created through simulation and in real subject motion. Future work will involve testing the robustness and performance of the method on motion corrupted data using a full 3D motion model.

Acknowledgements

No acknowledgement found.

References

1. Andre JB, Bresnahan BW, Mossa-Basha M, et al. Toward quantifying the prevalence, severity, and cost associated with patient motion during clinical MR examinations. J Am Coll Radiol. 2015;12(7):689-695. doi:10.1016/j.jacr.2015.03.007.

2. Zaitsev M, Maclaren J, Herbst M. Motion artifacts in MRI: A complex problem with many partial solutions. J Magn Reson Imaging. 2015;42(4):887-901. doi:10.1002/jmri.24850.

3. Loktyushin A, Nickisch H, Pohmann R, Schölkopf B. Blind retrospective motion correction of MR images. Magn Reson Med. 2013;70(6):1608-1618. doi:10.1002/mrm.24615.

4. Cordero-Grande L, Hughes EJ, Hutter J, Price AN, Hajnal J V. Three-dimensional motion corrected sensitivity encoding reconstruction for multi-shot multi-slice MRI: Application to neonatal brain imaging. Magn Reson Med. 2018;79(3):1365-1376. doi:10.1002/mrm.26796.

5. Haskell MW, Cauley SF, Wald LL. TArgeted Motion Estimation and Reduction (TAMER): Data Consistency Based Motion Mitigation for MRI Using a Reduced Model Joint Optimization. IEEE Trans Med Imaging. 2018;37(5):1253-1265. doi:10.1109/TMI.2018.2791482.

6. Hammernik K, Klatzer T, Kobler E, et al. Learning a variational network for reconstruction of accelerated MRI data. Magn Reson Med. 2018;79(6):3055-3071. doi:10.1002/mrm.26977.

7. Pruessmann KP, Weiger M, Scheidegger MB, Boesiger P. SENSE: sensitivity encoding for fast MRI. Magn Reson Med. 1999;42(5):952-962. http://www.ncbi.nlm.nih.gov/pubmed/10542355.

8. Bilgic B, Cauley SF, Chatnuntawech I, et al. Combining MR Physics and Machine Learning to Tackle Intractable Problems. In: Proceedings of the 26th Annual Meeting of ISMRM, Paris. ; 2018:3374.

9. Batchelor PG, Atkinson D, Irarrazaval P, Hill DLG, Hajnal J, Larkman D. Matrix description of general motion correction applied to multishot images. Magn Reson Med. 2005;54(5):1273-1280. doi:10.1002/mrm.20656.

10. Cordero-Grande L, Teixeira RPAG, Hughes EJ, Hutter J, Price AN, Hajnal J V. Sensitivity Encoding for Aligned Multishot Magnetic Resonance Reconstruction. IEEE Trans Comput Imaging. 2016;2(3):266-280. doi:10.1109/TCI.2016.2557069.

11. Hennig J, Nauerth A, Friedburg H. RARE imaging: a fast imaging method for clinical MR. Magn Reson Med. 1986;3(6):823-833. doi:10.1002/mrm.1910030602.

12. Odille F, Vuissoz P-A, Marie P-Y, Felblinger J. Generalized Reconstruction by Inversion of Coupled Systems (GRICS) applied to free-breathing MRI. Magn Reson Med. 2008;60(1):146-157. doi:10.1002/mrm.21623.

Figures

NAMER method overview. First, a motion corrupted image is reconstructed from the multicoil data, assuming no motion occurred. Next, motion mitigation is performed by looping through three steps: (1) remove motion artifacts in image space by passing the image through the motion detecting CNN, (2) search for the motion parameters by minimizing the data consistency error of a motion-inclusive forward model, using voxel values from the CNN image, and (3) reconstruct the full image volume using the motion-inclusive forward model and position coordinated from step (2).

Convolutional neural network for motion artifact detection. A motion corrupted image is input to a 27-layer patch-based CNN consisting of convolutional layers, batch normalization, and ReLU nonlinearities. The network outputs the image artifacts, which can be subtracted from the input image to arrive at a motion mitigated image.

CNN artifact mitigation across whole brain. (Left) Representative slices of CNN motion artifact mitigation across the brain volume for a simulated motion example. The CNN used was trained on four healthy subjects and applied to a fifth healthy subject. Bottom right shows the image space RMSE compared to ground truth image. Ringing artifacts are reduced, yet some still remain and slight blurring is introduced. (Right) For all slices in the brain volume, the image space RMSE decreased after the CNN, with an average improvement of 3.9%.

NAMER convergence compared to non-ML method in simulated motion data. A. Convergence for a conventional (no CNN) method compared to NAMER. A single iteration includes all three NAMER steps, or just the second and third steps for the non-ML version. NAMER converges within 8 steps, while the non-ML method takes 50+ steps to converge. B. Reconstruction of motion corrupted k-space data using motion values equal to zero (left) and ground truth motion parameters (right). C. Image reconstruction for non-CNN and NAMER methods at the 1st and 5th iterations. Blue values are % data consistency error.

NAMER artifact mitigation on real motion cases. The NAMER reconstruction was applied to two subjects, both with anatomy unseen by the CNN during training. Left column: Original motion corrupted images. Middle column: NAMER motion mitigated images. Right column: Difference image between the original and NAMER images, scaled at 2x relative to the other columns. Position coordinates found during the NAMER algorithm are shown at right, with trajectories similar to those expected, based on instructions given to subjects.

Proc. Intl. Soc. Mag. Reson. Med. 27 (2019)
0074