1840

Data-Efficient Uncertainty Quantification for Radial Cardiac Cine MR Image Reconstruction
Sherine Brahma1, Tobias Schäffter1,2,3, Christoph Kolbitsch1,3, and Andreas Kofler1
1Physikalisch-Technische Bundesanstalt, Braunschweig and Berlin, Germany, 2Department of Biomedical Engineering, Technical University of Berlin, Berlin, Germany, 3School of Imaging Sciences and Biomedical Engineering, King’s College London, London, United Kingdom

Synopsis

Due to the black-box nature of Deep Learning (DL) algorithms, uncertainty quantification (UQ) is a promising approach to assess their risk in medical applications. However, UQ is challenging in imaging techniques like non-Cartesian multi-coil CINE MRI because the data is high-dimensional, and the acquisition process is computationally demanding. In this work, i) we propose to utilize spatio-temporal (ST) networks, demonstrating efficient UQ in a high-dimensional setting, and ii) we show a reduction in uncertainty by adopting the forward model of radial multi-coil 2D cine MR in the reconstruction process. UQ is performed using MCMC dropout with additional aleatoric loss terms.

Introduction

Despite achieving excellent image quality, adopting DL models for MR reconstructions in clinical routine is a concern because of its black-box nature. In such cases, UQ has the potential to provide feedback where the results provided by DL cannot necessarily be trusted.
UQ involves learning a posterior distribution, which is a hard problem in dynamic MR reconstruction because it deals with high-dimensional (space and time) data. Uncertainties can primarily stem from i) the inherent noise present in the output, called the aleatoric uncertainty, and ii) the uncertainties in model parameters, called the epistemic uncertainty [1]. Previous work demonstrated the usage of MCMC dropout with aleatoric loss terms to characterize uncertainty in CINE MRI reconstruction [2]. However, to the best of our knowledge, previous works only considered the computationally less demanding Cartesian single-coil acquisition case. Here, we address the problem of uncertainty quantification for a 2D radial multi-coil cine MR reconstruction problem. Further, we exploit the benefits of ST networks [3,4] at a high-dimensional setting.

Methods

Our goal is to obtain the unknown MRI image $$$\mathbf{x} \in \mathbb{C}^{N_{x} \cdot N_{y}\cdot N_{t} }$$$ from the following ill-posed problem
$$\mathbf{y}_{I} = \mathbf{A}_{I}\mathbf{x} + \mathbf{e} \tag{1},$$
where the operator $$$\mathbf{A}_{I}$$$ contains the radial k-space trajectories and the multi-coil sensitivity maps information. The measured undersampled k-space data is denoted by $$$\mathbf{{y}_{I}}$$$ and is corrupted by random Gaussian noise $$$\mathbf{e}$$$.

End-to-end network: It incorporates acquisition information into the reconstruction process by a three steps training procedure [3] illustrated in Fig. 1 (bottom). The final data-consistent solution $$$\mathbf{x}_{\mathrm{CG}}$$$ is obtained by applying a conjugate gradient algorithm to minimize the problem below, where the network output $$$\mathbf{x}_{\mathrm{CNN}}$$$ serves as a prior.
$$\mathbf{x}_{\mathrm{CG}}:= \underset{\mathbf{x}}{\displaystyle{\arg \min}} \; ||\mathbf{W}^{\frac{1}{2}}(\mathbf{A}_{I}\mathbf{x}-\mathbf{y}_{I})||_{2}^{2}+\lambda||\mathbf{x}-\mathbf{x}_{\mathrm{CNN}}||_{2}^{2} \tag{2}.$$
Here, $$$ \mathbf{W} $$$ is the density compensation matrix.

Spatio-temporal network: A key element of our work is to adopt an ST-Unet (referred here as XTYT-Unet) that takes inputs along the ST perspective. In [4], the authors found that it had the best performance among all 2D-networks and performed comparably well as a 3D-UNet [5], but with a reduced computationally complexity.

Uncertainty quantification: The abbreviations E, C, and K describing the architecture of $$$u_{\theta}$$$ is explained in Fig. 1 (top). Dropout layers with a dropout rate $$$p$$$ are applied before every convolutional block. The following loss function is minimized during training [1,6]:
$$\mathcal{L}(\boldsymbol{\theta})=-\frac{1}{D}\sum_{d=1}^D (\frac{1}{2 (\boldsymbol{\sigma}^{(d)})^2}||\mathbf{x}_{f}^{(d)}-\mathbf{x}_{\scriptscriptstyle{CG}}^{(d)}||_{2}^{2}+\frac{1}{2}\log((\boldsymbol{\sigma}^{(d)})^2)+\frac{1-p}{2D}||\boldsymbol{\theta}||_{2}^{2} \tag{3}.$$
Here, $$$D$$$ is the total pixel number, $$$\sigma$$$ is the aleatoric uncertainty, $$$x_{f}$$$ is the ground truth, and $$$ \theta $$$ denotes network parameters.
The total uncertainty is estimated as the square root of the predictive variance using $$$S$$$ samples-
$$\mathrm{Var}(\mathbf{x}_{\mathrm{CG}})\approx \underbrace{\frac{1}{S}\sum_{s=1}^S (\mathbf{x}_{\mathrm{CG}}^{(s)})^2-(\frac{1}{S}\sum_{s=1}^S \mathbf{x}_{\mathrm{CG}}^{(s)})^2}_{\text{epistemic}}+\underbrace{\frac{1}{S}\sum_{s=1}^S(\boldsymbol{\sigma}^{(s)})^2}_{\text{aleatoric}} \tag{4},$$
where the first two terms correspond to the epistemic uncertainty and the last one to the aleatoric uncertainty.

Dataset and evaluation: Our study uses a dataset of 216 complex-valued cine MRI images with dimensions 320x320x30 (2D space + time dimension) split into 144/36/36 images for training, validation, and testing respectively [3]. The reconstruction quality for the test-set is quantified using NRMSE, PSNR, and SSIM [7].

Result

Fig. 2 illustrates the NRMSE trend of XTYT and 3D-Unets during different training phases. The trend for 3D-Unet E3C2K16, trained with a mini-batch size of 2 (mb=2), fails to converge during pretraining. However, when the training images are cropped to 160x160x30, and a lower number of filters in the network are used to be able to fit an mb=16 in the memory, the 3D-Unet E3C2K8 converges. Cropped image size of 160x160x30, however, cannot be used for finetuning because the forward operator requires the entire image whose k-space data is sampled. The training and validation loss of XTYT-Unet E3C2K16 declines further with finetuning after pretraining. Fig. 3 shows that finetuning enhances the RMSE, PSNR, and SSIM values of a pretrained XTYT-Unet E3C2K16. A visual inspection (Fig. 4) of an example slice illustrates the decrease in error and uncertainties for the same finetuned network.

Discussion

There is a good possibility that mb=2 is not sufficient to capture the data uncertainty $$$\sigma$$$ in the output for the 3D-Unet E3C2K16. Hence, its loss fails to converge. This hypothesis is evidenced by the convergence of 3D-Unet E3C2K8 when the mini-batch size is increased to 16. By comparison, the XTYT-Unet E3C2K16 allows for more effective $$$\sigma$$$ learning since slices in this perspective appear to be similar and simpler in structure, holding the observation of [4].
Finally, Fig. 3 and Fig. 4 suggest that including the forward model in the finetuning stage not only reduces the estimation error but also helps decrease the overall uncertainty. The reduction in epistemic uncertainty can be attributed to the fact that the forward model incorporates useful knowledge about the reconstruction process, while aleatoric uncertainty decreases because of less statistical noise in the network output due to improved modeling. Regions of high uncertainty correlate with the error and could be useful input parameters for further diagnosis or even automatic segmentation.

Conclusion

Here, we presented a network that allows for efficient UQ even for a high-dimensional image reconstruction problem such as in non-Cartesian multi-coil CINE MRI. Furthermore, experimental results favor the inclusion of the forward model in the reconstruction process as it decreases the estimation error as well as the uncertainty.

Acknowledgements

We gratefully acknowledge funding from the German Research Foundation (GRK2260, BIOQIC)

References

[1] Kendall, Alex, and Yarin Gal. "What uncertainties do we need in bayesian deep learning for computer vision?." arXiv preprint arXiv:1703.04977 (2017).

[2] Schlemper, Jo, et al. "Bayesian deep learning for accelerated MR image reconstruction." International Workshop on Machine Learning for Medical Image Reconstruction. Springer, Cham, 2018.

[3] Kofler, Andreas, et al. "An end‐to‐end‐trainable iterative network architecture for accelerated radial multi‐coil 2D cine MR image reconstruction." Medical Physics 48.5 (2021): 2412-2425.

[4] Kofler, Andreas, et al. "Spatio-temporal deep learning-based undersampling artefact reduction for 2D radial cine MRI with limited training data." IEEE transactions on medical imaging 39.3 (2019): 703-717.

[5] Hauptmann, Andreas, et al. "Real‐time cardiovascular MR with spatio‐temporal artifact suppression using deep learning–proof of concept in congenital heart disease." Magnetic resonance in medicine 81.2 (2019): 1143-1156.

[6] Gal, Yarin, and Zoubin Ghahramani. "Dropout as a bayesian approximation: Representing model uncertainty in deep learning." international conference on machine learning. PMLR, 2016.

[7] Wang, Zhou, et al. "Image quality assessment: from error visibility to structural similarity." IEEE transactions on image processing 13.4 (2004): 600-612.

Figures

Figure 1: The sequence of training end-to-end network is split into three parts as shown in the numbered orange boxes. An initial estimate $$$\mathbf{{x}_{I}}$$$ is used as an input for pretraining a spatio-temporal network $$$u_{\theta}$$$. This network reduces the undersampling artifacts in spatio-temporal domain rather than in the 2D spatial domain. In the final step, we include $$$\mathbf{A}_{I}$$$ in the conjugate gradient minimization problem in Eqn. (2) and finetune the model end-to-end.

Figure 2: On the left image, the key point to note here is that, unlike the XTYT-Unet (black curves), the 3D-Unet (red curves) fails to converge even when both networks are trained with the same mini-batch size of 2. On the right, the plot indicates that to make the 3D-Unet (blue curves) converge, we require a high enough mini-batch size (16 is used here), making an expensive tradeoff in memory requirement.

Figure 3: This table shows a comparison of the test-set performance measures between the pretrained and the finetuned network. As we can see, the finetuned network, which incorporates the forward model, has a lower NRMSE, and higher PSNR and SSIM.

Figure 4: The visual inspection of a test slice is shown above. Each 2D spatial view is accompanied by an ST view on the left, sliced along the dashed red line shown in (a). (a) is the ground truth, and the networks are tested to reduce the undersampling artifacts of (g). The arrows compare the results of the pretrained network (b, c, d, e, f) and the finetuned network (h, i, j, k, l) by highlighting the respective regions in a column-wise manner. As compared to the pretrained network, we see a decrease in epistemic, aleatoric, and total uncertainty, and error for the finetuned network.

Proc. Intl. Soc. Mag. Reson. Med. 30 (2022)
1840
DOI: https://doi.org/10.58530/2022/1840