1953

Joint Neural Network for Fast Retrospective Rigid Motion Correction of Accelerated Segmented Multislice MRI
Nalini M Singh1,2, Malte Hoffmann3,4, Daniel C Moyer1, Ikbeom Jang3,4, Lina Chen5, Marcio Aloisio Bezerra Cavalcanti Rockenbach5, Arnaud Guidon6, Iman Aganj3,4, Jayashree Kalpathy-Cramer3,4,5, Elfar Adalsteinsson2,7,8, Bruce Fischl2,3,4, Adrian V Dalca1,3,4, Polina Golland*1,7,8, and Robert Frost*3,4
1Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, MA, United States, 2Harvard-MIT Health Sciences and Technology, Massachusetts Institute of Technology, Cambridge, MA, United States, 3Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital, Charlestown, MA, United States, 4Department of Radiology, Harvard Medical School, Boston, MA, United States, 5MGH & BWH Center for Clinical Data Science, Mass General Brigham, Boston, MA, United States, 6GE Healthcare, Boston, MA, United States, 7Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology, Cambridge, MA, United States, 8Institute for Medical Engineering and Science, Massachusetts Institute of Technology, Cambridge, MA, United States

Synopsis

We demonstrate a deep learning approach for fast retrospective intraslice rigid motion correction in segmented multislice MRI. The proposed neural network architecture combines convolutions on frequency and image space representations of the acquired data to produce high quality reconstructions. Unlike many prior techniques, our method does not require auxiliary information on the subject head motion during the scan. The resulting reconstruction procedure is more accurate and is an order of magnitude faster than GRAPPA. Our work offers the first step toward fast motion correction in any setting with substantial, unpredictable, challenging to track motion.

Authorship Information

* Denotes equal contribution.

Introduction

Motion frequently corrupts MRI acquisitions, degrading image quality or necessitating repeated scans1. We demonstrate a deep learning technique for fast retrospective intraslice rigid motion correction in 2D fast spin echo (FSE) brain MRI. In contrast to the existing deep learning retrospective motion correction methods,2–7 we extend our previously demonstrated architecture that interleaves frequency and image space convolutions. Joint reconstruction outperforms pure image space architectures in the single-coil accelerated acquisition setting8. Here we demonstrate advantages of the joint architecture for multicoil, multishot acquisitions.

We assume a quasistatic motion model in which the object remains stationary during individual shots and moves via 2D rotations and translations between shots. We defer modeling through-slice motion effects for future work.

Methods

Data
We demonstrate our approach on k-space data from 2D FLAIR FSE brain MRI scans (3T GE Signa Premier, 48-channel head coil, TR=10000ms, TE=118ms, FOV=260x260mm2, acquisition matrix=416x300, slice thickness=5mm, slice spacing=1mm) acquired at our hospital under an approved IRB protocol. We treat the acquired data as ground truth and simulate motion artifacts during training and testing. We split the dataset into 1064 training 2D images (76 subjects), 396 validation images (18 subjects), and 362 test images (17 subjects), with no subject overlap among sets.

Network Architectures and Training
Our joint network learns combinations of frequency and image space features at each layer (Figure 1). The network’s input is an 88-channel image composed of the real and imaginary components of the k-space data from 44 receive coils (4 neck channels from the 48-channel head coil were discarded). The input is processed via 6 successive interleaved layers followed by a single convolution, yielding an 88-channel k-space output converted to the reconstructed image via the inverse Fourier transform and root-sum-of-squares coil combination. We train the network using the structural similarity index measure (SSIM) loss function that compares the reconstruction with the ground-truth, motion-free image.

Motion Simulation
Figure 2 illustrates our motion simulation process. For each image, we select at random a shot at which motion occurs, sample the translation parameters $$$\Delta x$$$ and $$$\Delta y$$$ from a uniform distribution over the range [-10mm,+10mm], and sample the rotation parameter $$$\Delta \theta$$$ from a uniform distribution over the range [-10o,10o].

We apply GRAPPA9 to the acquired motion-free k-space data $$$F_O$$$, followed by the inverse Fourier transform and root-sum-of-squares coil combination to reconstruct the motion-free image $$$I_O$$$. We estimate coil sensitivity profiles $$$S_i$$$ using ESPIRiT10 with learned parameter estimation.11

To synthesize the post-movement image $$$I_M$$$, we apply the sampled translation and rotation to $$$I_O$$$. We synthesize the corresponding k-space data that would be acquired by coil $$$i$$$ as $$$F_{M,i} = P\mathcal{F}S_i I_M$$$, where $$$P$$$ is the undersampling operator and $$$\mathcal{F}$$$ is the Fourier transform matrix. Finally, we generate the motion-corrupted k-space data $$$F_C$$$ by interleaving the k-space lines from $$$F_O$$$ and $$$F_M$$$ corresponding to the pre- and post-motion shots, respectively.

We normalize each input and output pair by dividing by the maximum intensity in the corrupted image. Our network reconstructs a normalized version of whichever of $$$F_O$$$ or $$$F_M$$$ corresponds to the position of the subject during acquisition of the central k-space line of $$$F_C$$$.

Implementation Details
We compare the proposed network with baseline architectures composed solely of frequency space convolutions or image space convolutions with twice as many layers as the joint model, such that the total number of convolutions is consistent among all models. We train all networks using the Adam optimizer (learning rate 1e-3) with a batch size of 6. We form each batch using images from the same volume to maintain constant k-space matrix size within a batch. We omit batch normalization when training pure image space networks, as we empirically found it hurt performance. We tune the hyperparameters on the validation set and use the test set for the final evaluation. The networks take 1-3 days to train and 400 milliseconds to run on a new image on an NVIDIA RTX 8000 GPU.

Results and Discussion

Our method improves over GRAPPA reconstruction of the motion-corrupted k-space data for the majority of subjects and in particular improves images for which GRAPPA performs poorly (Figure 3a). The reconstruction speed is an order of magnitude faster than that of GRAPPA. Our reconstructions outperform the single-space networks for nearly every subject. We also examine reconstruction quality when motion occurs at different shots during the acquisition (Figure 3b) and for varying motion magnitude (Figures 3c-e). Figure 4 illustrates visually that the joint architecture produces sharper, more accurate reconstructions than the GRAPPA baseline and both single-space architectures. Figure 5 shows qualitatively that our method outperforms GRAPPA for severe motions where GRAPPA performs poorly.

Conclusions

We demonstrate a deep learning approach for correcting intraslice rigid motion artifacts in 2D FSE FLAIR MRI and characterize the motion range for which our method succeeds. Future work will extend our approach to model 3D, through-slice motion artifacts and nonrigid motion. Further, we will investigate optimization strategies to ensure consistency of the output with estimated or measured motion parameters. Our technique is a step toward flexible deep learning motion correction for a wide variety of imaging scenarios affected by significant subject motion.

Acknowledgements

Research reported in this abstract was supported in part by GE Healthcare, NIH NIBIB (5T32EB1680, P41EB015896, 1R01EB023281, R01EB006758, R21EB018907, R01EB019956, R01EB017337, P41EB03000, R21EB029641), NIBIB NAC (P41EB015902), NICHD (U01HD087211, K99 HD101553), NIA (1R56AG064027, 1R01AG064027, 5R01AG008122, R01AG016495, 1R01AG070988, RF1AG068261), NIMH (R01 MH123195, R01 MH121885, 1RF1MH123195), NINDS (R01NS0525851, R21NS072652, R01NS070963, R01NS083534, 5U01NS086625, 5U24NS10059103, R01NS105820), and the Blueprint for Neuroscience Research (5U01-MH093765), part of the multi-institutional Human Connectome Project. Additional support was provided by the BRAIN Initiative Cell Census Network grant U01MH117023, and a Google PhD Fellowship.

References

1. Andre, J. B. et al. Toward Quantifying the Prevalence, Severity, and Cost Associated With Patient Motion During Clinical MR Examinations. J. Am. Coll. Radiol. 12, 689–695 (2015).

2. Duffy, B. A. et al. Retrospective correction of motion artifact affected structural MRI images using deep learning of simulated motion. (2018).

3. Haskell, M. W. et al. Network Accelerated Motion Estimation and Reduction (NAMER): Convolutional neural network guided retrospective motion correction using a separable motion model. Magn. Reson. Med. 82, 1452–1461 (2019).

4. Johnson, P. & Drangova, M. Conditional generative adversarial network for three-dimensional rigid-body motion correction in MRI. in (2019).

5. Küstner, T. et al. Retrospective correction of motion-affected MR images using deep learning frameworks. Magn. Reson. Med. 82, 1527–1540 (2019).

6. Oksuz, I. et al. Detection and Correction of Cardiac MRI Motion Artefacts During Reconstruction from k-space. in Medical Image Computing and Computer Assisted Intervention – MICCAI 2019 695–703 (Springer International Publishing, 2019).

7. Usman, M., Latif, S., Asim, M., Lee, B.-D. & Qadir, J. Retrospective Motion Correction in Multishot MRI using Generative Adversarial Network. Sci. Rep. 10, 4786 (2020).

8. Singh, N. M., Iglesias, J. E., Adalsteinsson, E., Dalca, A. V. & Golland, P. Joint Frequency- and Image-Space Learning for Fourier Imaging. arXiv [cs.CV] (2020).

9. Griswold, M. A. et al. Generalized autocalibrating partially parallel acquisitions (GRAPPA). Magn. Reson. Med. 47, 1202–1210 (2002).

10. Uecker, M. et al. ESPIRiT--an eigenvalue approach to autocalibrating parallel MRI: where SENSE meets GRAPPA. Magn. Reson. Med. 71, 990–1001 (2014).

11. Iyer, S., Ong, F., Setsompop, K., Doneva, M. & Lustig, M. SURE-based automatic parameter selection for ESPIRiT calibration. Magn. Reson. Med. 84, 3423–3437 (2020).

Figures

Proposed network architecture. Our network takes as input 44-coil undersampled motion-corrupted k-space data represented as 88 channels with real and imaginary components for each coil and produces motion-free, fully-sampled k-space data in the same format. We employ the inverse Fourier transform and root-sum-of-squares coil combination to generate the output image. Each interleaved layer learns linear combinations of the previous layers' frequency and image space outputs $$$u_n$$$ and $$$v_n$$$ and applies convolutions in both spaces.

Motion simulation. We estimate an image $$$I_O$$$ by applying the inverse Fourier transform and root-sum-of-squares coil combination to a GRAPPA-reconstructed version $$$F_O$$$ of acquired motion-free k-space data. We rotate and translate $$$I_O$$$ to generate moved image $$$I_M$$$, which is multiplied by coil sensitivities $$$S_i$$$ and Fourier transformed to simulate moved k-space $$$F_M$$$. Lines from $$$F_M$$$ and $$$F_O$$$ are combined to form simulated k-space $$$F_C$$$ that captures motion from $$$I_O$$$ to $$$I_M$$$ between acquisition of subsequent shots.

Quantitative Results. (a) Subject-specific SSIM (sorted by GRAPPA reconstruction value) for our method and baselines. The joint network outperforms all other methods for the vast majority of cases. (b) Reconstructed SSIM as a function of the shot at which motion occurred. Motion right before shots 3 and 4, which heavily sample low frequencies in k-space, results in lower reconstruction quality. (c-e) SSIM for a range of motion parameters, with all other parameters held constant. The dropoff in reconstruction quality for larger motion is lower for our approach than the baselines.

Example Reconstruction Results. From top to bottom: Image reconstructions, zoomed-in image patches, patchwise SSIM maps, and k-space for a single coil. The joint network produces sharper images and preserves high-frequency components (e.g., skull details) better than GRAPPA and single-space networks.

Reconstruction Results for Varying Motion Parameters. Example reconstructions of data corrupted by small and large rotations and translations. The joint network produces sharper, more accurate images than GRAPPA, particularly for larger motions.

Proc. Intl. Soc. Mag. Reson. Med. 30 (2022)
1953
DOI: https://doi.org/10.58530/2022/1953