2216

Rapid myocardial perfusion MRI reconstruction using deep learning networks
Eric Kenneth Gibbons1,2, Ye Tian3, Qi Huang4, Akshay Chaudhari5, and Edward DiBella2,4
1Electrical and Computer Engineering, Weber State University, Ogden, UT, United States, 2Radiology and Imaging Sciences, University of Utah, Salt Lake City, UT, United States, 3Physics, University of Utah, Salt Lake City, UT, United States, 4Biomedical Engineering, University of Utah, Salt Lake City, UT, United States, 5Radiology, Stanford University, Stanford, CA, United States

Synopsis

Current acquisition strategies in cardiac perfusion MRI rely on non-uniform sampling that is highly undersampled in spatial and temporal domains. While iterative reconstruction methods are able to reconstruct such data reasonably well, reconstruction speeds are prohibitively long. This abstract applies novel deep learning approaches to accelerate reconstruction speeds relative to iterative algorithms with comparable image quality. Validation is performed through the calculation of a perfusion index.

Introduction

Rapid acquisition and reconstruction techniques are critical for cardiac perfusion MRI. However, the current state of the art methods, such as constrained reconstruction, are able to provide high-quality images from sparsely sampled k-space, though at the cost of slower reconstruction speeds[1]. Such algorithms will take minutes to reconstruct a full dataset, which consists of 2-4 image slices and 70-100 image frames, even when such algorithms are implemented using cutting edge computing techniques. This is problematic from a clinical perspective as it precludes immediate feedback for the scanner operator as to whether the scan quality is sufficient.

Advances in neural networks have provided tools to achieve similar reconstruction quality seen in constrained reconstruction methods, but are computationally much more efficient after the network is sufficiently trained[2]. In this work, we apply deep learning techniques to radially sampled cardiac perfusion MRI datasets and compare reconstruction speeds, image quality, and calculated perfusion index.

Methods

Cardiac perfusion datasets were acquired from thirty participant subjects. All MRI data were acquired using either a Siemens 3T Trio or a 3T Verio scanner using a 12-element phased-array coil. Dynamic contrast-enhanced myocardial perfusion imaging was done first at rest, then during adenosine infusion according to [3]. The contrast was injected approximately 3 minutes after the start of the adenosine infusion and approximately 90 seconds after the regadenoson injection. A saturation recovery radial turboFLASH pulse sequence was used. Seventy-two rays were acquired after each saturation pulse. Additional scan parameters were TR of 2.6 ms, a TE of 1.14 ms, flip angle of 14deg. The voxel size was 2.3x2.3x8mm. Each scan included 2-4 slices acquired for each heartbeat for approximately 1 minute during free breathing.

Data were reconstructed using a neural network architecture operating in the image space. The raw radial k-space data were retrospectively undersampled to 24 rays in a quasi-golden angle fashion. This k-space data were then gridded and transformed to image space using a NUFFT. Magnitude images were fed into the network, which consisted primarily of a U-net[4] architecture. A standard input and output training pair consisted of one slice. On the input, 15 consecutive time frames of the same slice were using and treated as separate channels. The middle channel (in this case, the eighth channel) was considered to represent the target time-frame. On the output, a single channel final image was produced. Fig.1 includes the input/output relationship of this network, the network architecture, and sampling pattern. Because each participant was scanned multiple times with multiple injections, and each scan included multiple slices, and each slice was scanned with 60-100 time frames, 10700 training samples along with 2300 validation samples were used to train the network.

Data from three separate study participants (for a total of ten total injections) were withheld from model training and validation as test data. These images processed by the neural network were benchmarked by comparing them to images reconstructed using standard iterative techniques using the full 72 rays (considered here to be the reference) and the subsampled 24 rays.

To calculate a perfusion index, images were first registered for inter-frame motion, and then a region of interest in the blood pool and tissue borders were defined manually. Time curves from the blood region and six tissue regions per slice were then converted to approximate gadolinium concentrations and fit with a compartment model to give a perfusion index.

Results

Fig. 2 shows example output images from the fully sampled reference and the undersampled 24 ray STCR and deep learning reconstructions. In the difference images between the 24 ray deep learning images and the reference, anatomical structures do not appear to be present, particularly in comparison with the undersampled STCR case. Looking at the contrast time curves in the blood pool in Fig. 3, the contrast in the images generated by the deep learning approach match with the reference closely. Additionally, the entire reconstruction process (from raw k-space data to final image) took 12-16 seconds using the deep learning methods. By contrast, the iterative approach took 3-4 minutes based on the number of frames and slices.

Fig. 4 shows the mean calculated perfusion index for each of the contrast injections for each of the test participants. These trends suggest that the deviations from the reference for the deep learning images are in line with what we see with established iterative reconstruction techniques.

Discussion and Conclusion

The results presented here show that a deep learning approach can give comparable results to images generated by iterative algorithms at a much lower computational cost. However, pushing beyond the results established here likely will require more training data as well as more exotic network architectures. Incorporating adversarial loss has been demonstrated to improve neural network performance in these reconstruction tasks. Additional validation is an important next step.

We have demonstrated a rapid method of generating high fidelity images from highly undersampled radial cardiac perfusion k-space datasets. This deep learning approach achieves considerably faster reconstruction times than iterative reconstruction methods.

Acknowledgements

No acknowledgement found.

References

[1] Tian, Ye, et al. "Evaluation of pre‐reconstruction interpolation methods for iterative reconstruction of radial k‐space data." Medical Physics 44.8 (2017): 4025-4034.

[2] Mardani, Morteza, et al. "Deep generative adversarial neural networks for compressive sensing MRI." IEEE Transactions on Medical Imaging 38.1 (2018): 167-179.

[3] DiBella, Edward VR, et al. "The effect of obesity on regadenoson-induced myocardial hyperemia: a quantitative magnetic resonance imaging study." The International Journal of Cardiovascular Imaging 28.6 (2012): 1435-1444.

[4] Ronneberger, Olaf, Philipp Fischer, and Thomas Brox. "U-net: Convolutional networks for biomedical image segmentation." International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, Cham, 2015.

Figures

Fig. 1: (a) The inputs and outputs of the network. The inputs consist of a 2D image with 15 channels. Each channel represents an adjacent time-frame. The eighth (i.e., the central) channel is used in the residual connection. After the image is processed by the network, the central 50% of the image is cropped for display. (b) The proposed network follows a U-net architecture. A 15-channel input (each channel corresponding to a different frame number) and the middle time frame channel is used as the residual image. (c) This approach relies on quasi-golden angle sampling.

A time-series reconstruction is given for 72 ray STCR (considered to be the reference standard here), 24 ray STCR, and the 24 ray deep learning reconstructions. Time frames were chosen to show the contrast as it enters the blood pool and exits into the surrounding myocardial tissue. Difference images for both 24 ray cases are also given.

Fig. 3: A mean pixel value plot is given for the blood pool region-of-interest for the 72 ray STCR (considered to be the reference standard here), 24 ray STCR, and 24 ray deep learning reconstruction. In (a) and (c), the region-of-interest is highlighted in red in the blood pool and the myocardium, respectively. In (b) and (d), the time curves show that neither 24 ray reconstruction results deviate too strongly from the reference.

Perfusion index values compared to each of the three methods. The index was calculated over six regions over the myocardium and is averaged. Each scan participant was given 3-4 contrast injections, and each individual injection is displayed here.

Proc. Intl. Soc. Mag. Reson. Med. 28 (2020)
2216