3885

Accelerated MRI using Dual-domain Transformer-Based Reconstruction and Learning-based Undersampling
Guan Qiu Hong1,2,3, William Morley2, Matthew Wan2, Yuan Tao Wei1,2,3, Yang Su2, and Hai-Ling Margaret Cheng1,2,3
1Institute of Biomedical Engineering, University of Toronto, Toronto, ON, Canada, 2The Edward S. Rogers Sr. Department of Electrical and Computer Engineering, University of Toronto, Toronto, ON, Canada, 3Translational Biology & Engineering Program, Ted Rogers Centre for Heart Research, Toronto, ON, Canada

Synopsis

Keywords: Machine Learning/Artificial Intelligence, Image Reconstruction

Deep learning architectures such as Convolutional Neural Networks (CNNs) are widely used alongside fixed k-space undersampling for reconstructing highly undersampled MRI k-space data. However, CNN-based architectures may perform sub-optimally due to their limited ability to capture long range dependencies, and fixed undersampling patterns may not be amenable to optimal reconstruction. We propose dual-domain (image and k-space), transformer-based reconstruction architectures paired with learning-based undersampling to reconstruct undersampled k-space data. Experimental results demonstrate improved reconstruction quality compared to CNN-based (UNet) architecture across 5x to 100x acceleration factors; dual domain outperformed single domain reconstructions.

Introduction

MRI is an inherently slow modality, and despite decades of progress in acceleration (e.g. parallel imaging1-4 and compressed sensing5), acceleration factors remain modest. Recently, there has been great interest in exploring deep learning for reconstructing images from greatly undersampled MRI data. In this work, we present a dual-domain (i.e. image and k-space domain) reconstruction framework using transformers, leveraging their self-attention mechanism to perform image domain anti-aliasing, k-space interpolation, and combining the dual domain reconstruction outputs into a final image domain reconstruction.

Methods

An overview of our model is depicted in Figure 1, and detailed architecture of each sub-network is shown in Figure 2. The model accepts sets of five images at a time. Because cardiac CINE scans consist of a sequence of temporally coherent heart images, using sequences of images as input incorporates temporal information for more robust reconstructions. Let $$${\{x_i\in {\mathbb{R}}^{5\times d}\}}^n_{i=1}\ $$$ denote sets of five full-resolution input images, where $$$i$$$ refers to the $$${i}^{th}$$$ set, $$$d$$$ is the dimension of an image, and $$$n$$$ is total number of image sets. Our model solves the following optimization problem (1) $${\mathrm{min}}_{\theta ,\alpha ,\beta ,\gamma }\ \frac{1}{n}\sum^n_{i=1}{{\left\|C_{\gamma }\left(I_{\alpha }\left(U_{\theta }\left(x_i\right)\right),F^{-1}\left(K_{\beta }\left(F\left(U_{\theta }\left(x_i\right)\right)\right)\right)\right)-x_i\left[3\right]\right\|}^2_2}$$ where $$$U_{\theta }\left(\bullet \right)$$$ transforms images to k-space, performs undersampling parameterized by $$$\theta$$$, and transforms undersampled k-space back to aliased images. $$$I_{\alpha }\left(\bullet \right)$$$ denotes the image domain anti-aliasing function parameterized by $$$\alpha$$$; $$$K_{\beta }\left(\bullet \right)$$$ denotes the k-space interpolation function parameterized by $$$\beta$$$; $$$C_{\gamma }\left(\bullet \right)$$$ denotes the function that combines reconstructions from the two domains, parameterized by $$$\gamma$$$; $$$F\left(\bullet \right)$$$ is the 2D Fourier transform (applied separately to each image); $$$F^{-1}\left(\bullet \right)$$$ is the inverse 2D Fourier transform; $$$x_i\left[3\right]$$$ refers to the third (central) image from the sampled set; and $$${\left\|\bullet \right\|}_2$$$ denotes the L2 norm.
The $$$U_{\theta }$$$ function was implemented by the Undersampling network, which is a modification of the LOUPE network$$${}^{6}$$$ to allow 5 input images at a time. The $$$I_{\alpha }$$$, $$$K_{\beta }$$$, and $$$C_{\gamma }$$$ functions were implemented by the Image-domain transformer, k-Space transformer, and Combination transformer, respectively, all of which are modified versions of the UFormer$$${}^{7}$$$, shown in Figure 2. In the Image-domain transformer, given five temporally correlated and aliased images, the network must learn an encoding that generates one unaliased image that resembles the middle input image. Hence, the network accepts 10-channel input consisting of five complex-valued aliased images, and the output projection was modified to produce a 1-channel residual image. To further facilitate reconstruction of the middle input image, its magnitude image equivalent was added to the residual image to form the output of the Image transformer. Similar modifications were made to produce the k-Space transformer and the Combination transformer, which all encode high dimensional input data and generate reconstructed output.
Experiments were conducted on fully sampled 2D CINE scans acquired in the four-chamber view (UK Biobank). We used ADAM optimizer8 to minimize mean squared error (Eq. 1). We also compared transformer-based reconstruction networks with CNN-based reconstruction by replacing the Image, k-Space, and Combination UFormers with UNets. Each variation of the reconstruction network was connected to an undersampling network and trained from scratch to optimize end-to-end undersampling plus reconstruction and to allow a fair comparison. Reconstruction performance was assessed using the Structural Similarity Index Measure (SSIM).

Results and Discussion

We begin by examining undersampling patterns learned by our model. Figure 3 shows that the shape of the learned sampling pattern varies across the temporal sequence, which is expected because the model should minimize sampling redundancy across a temporally coherent sequence. The second and third sets capture low spatial frequency information, which is responsible for image contrast, while the first, fourth and fifth sets of sampled k-space capture high spatial frequency information, which is responsible for image details such as edges and small features. To visualize the sampling pattern more holistically, the combined mask is shown, which bears resemblance to variable density sampling masks that favor low frequency regions. Figure 4 depicts reconstructions from several variants of the reconstruction network: full (image-domain, k-space, combination) transformers, image-domain transformer only, k-Space transformer only, and comparisons against UNet CNN reconstructions. The full transformer model best preserved valve details, heterogeneity of the myocardium, and the anatomy of all heart structures, and suffered least from blurring artifacts. When reconstructions were performed using a single domain, the Image domain models (Image transformer only, and Image UNet only) outperformed the k-Space-only models, as they better recovered small anatomical details and had less ringing artifacts, seen as striations in the heart chambers. Finally, reconstruction quality, as measured using the SSIM metric, decreased with higher acceleration; however, the transformer-based model outperformed UNet-based model across all acceleration factors (Figure 5).

Conclusion

We proposed a dual domain, transformer-based framework for reconstructing undersampled MRI acquisitions. Our results demonstrated superior performance of the transformer architecture over conventional UNets, and dual-domain versus single-domain reconstructions.

Acknowledgements

This work was supported by the Natural Sciences and Research Council of Canada Discovery Grant [grant #2019-06137], Canadian Institutes of Health Research [grant #PJT 175131], and Canada Foundation for Innovation/Ontario Research Fund [grant #34038].

References

1. Griswold, M.A., Jakob, P.M., Heidemann, R.M., Nittka, M., Jellus, V., Wang, J., Kiefer, B., Haase, A., 2002. Generalized Autocalibrating Partially Parallel Acquisitions (GRAPPA). Magn Reson Med 47, 1202–1210. https://doi.org/10.1002/mrm.10171

2. Griswold, M.A., Jakob, P.M., Nittka, M., Goldfarb, J.W., Haase, A., 2000. Partially Parallel Imaging With Localized Sensitivities (PILS). Magn Reason Med 44, 602-609

3. Pruessmann, K.P., Weiger, M., Scheidegger, M.B., Boesiger, P., 1999. SENSE: Sensitivity Encoding for Fast MRI. Magn Reson Med. 42, 952-962

4. Sodickson, D.K., Manning, W.J., 1997. Simultaneous Acquisition of Spatial Harmonics (SMASH): Fast Imaging with Radiofrequency Coil Arrays. Magn Reson Med. 38, 591-603. doi: 10.1002/mrm.1910380414

5. Lustig, M., Donoho, D.L., Santos, J.M., Pauly, J.M., 2008. Compressed sensing MRI: A look at how CS can improve on current imaging techniques. IEEE Signal Process Mag. https://doi.org/10.1109/MSP.2007.914728

6. Bahadir, C.D., Wang, A.Q., Dalca, A. v., Sabuncu, M.R., 2020. Deep-Learning-Based Optimization of the Under-Sampling Pattern in MRI. IEEE Trans Comput Imaging 6, 1139– 1152. https://doi.org/10.1109/TCI.2020.3006727

7. Wang, Z., Cun, X., Bao, J., Zhou, W., Liu, J., Li, H., 2021a. Uformer: A General U-Shaped Transformer for Image Restoration. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 17683-17693

8. Kingma, D.P., Ba, J., 2014. Adam: A Method for Stochastic Optimization. arXiv preprint arXiv:1412.6980

Figures

Figure 1: Overall architecture of the model. The model consists of an undersampling network that learns the optimal undersampling pattern, followed by a series of networks to perform reconstruction.

Figure 2: Architecture of sub-networks in the model. Omitted sections of the transformer models are identical to original architecture from UFormer7. Each module’s output shape is listed as [Number of channels x Height x Width]. FFT = Fast Fourier Transform.

Figure 3: Learned undersampling masks using the full transformer model at an acceleration factor of R=20 (top row); corresponding aliased images (bottom row); and combined mask (upper-right).

Figure 4. Reconstructions from variations of the reconstruction network (R=20 acceleration). Top row shows full heart coverage; bottom row shows zoomed-in version of box bounded in blue.

Figure 5. Full transformer model vs. full U-Net model SSIM comparisons across a range of acceleration factors (5x, 10x, 20x, 50x, 100x).

Proc. Intl. Soc. Mag. Reson. Med. 31 (2023)
3885
DOI: https://doi.org/10.58530/2023/3885