3444

Bayesian sensitivity encoding enables parameter-free, highly accelerated joint multi-contrast reconstruction
Alexander Lin1, Demba Ba1, and Berkin Bilgic2,3,4
1Harvard University, Cambridge, MA, United States, 2Harvard-MIT Health Sciences and Technology, Massachusetts Institute of Technology, Cambridge, MA, United States, 3Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, United States, 4Department of Radiology, Harvard Medical School, Boston, MA, United States

Synopsis

To address limitations with classical SENSE algorithms, we propose Bayesian Sensitivity Encoding (Bayes-SENSE), which obviates the need to tune regularization penalties, provides variance maps that can quantify algorithmic uncertainty, and is easily extendable to multi-contrast reconstruction. Bayes-SENSE is a synergistic combination of SENSE and Bayesian CS (BCS). We adapt recent work accelerating BCS to develop an efficient and highly parallelizable inference algorithm for Bayes-SENSE based on the conjugate gradient (CG) method. We evaluate Bayes-SENSE in several undersampling settings with parallel imaging, and demonstrate that it outperforms L2-/L1-SENSE in terms of reconstruction error while also providing the aforementioned benefits.

Motivation

Sensitivity Encoding (SENSE) is an established parallel imaging method [1], and admits regularization terms that can lead to better reconstruction. For example, L2-SENSE algorithm adds an L2 (i.e. Tikhonov) penalty on the image [2], while L1-SENSE combines SENSE with compressed sensing (CS) to penalize the L1 norm in a sparsifying domain (e.g. Wavelets) [3, 4]. Despite their effectiveness, these regularized algorithms have certain limitations. They often require effort in tuning a regularization penalty, which may be difficult to identify without knowledge of the ground-truth image. In addition, they do not quantify model uncertainty, which can indicate where reconstruction errors are likely to occur.

Method

Bayes-SENSE reconstructs an image $$$x\in\mathbb{C}^N$$$ from undersampled k-space measurements $$$y\in\mathbb{C}^{N \cdot C / R}$$$, where $$$R$$$ is the undersampling factor and $$$C$$$ is the number of coils. Fig1 provides a diagram of the reconstruction.

Model: The basic SENSE model assumes that $$$y = Ex+\varepsilon$$$, where $$$\varepsilon$$$ is k-space noise and $$$E\in\mathbb{C}^{N\cdot C/R\times N}$$$ is the encoding matrix containing the sensitivity coils, Fourier transform, and undersampling masks. In Bayes-SENSE, we extend this model by combining it with BCS [5]. We assume that $$$x=Wz$$$, where $$$W\in\mathbb{C}^{N \times N}$$$ is an orthogonal transform (e.g. Wavelet basis) and $$$z\in\mathbb{C}^N$$$ is a vector of sparse coefficients. Let $$$\alpha\in\mathbb{R}^N$$$ be a vector of regularization penalties for each element of $$$z$$$. We impose a $$$\mathcal{N}(0, 1/\alpha_i)$$$ prior on each pair $$$(\text{Real}(z_i), \text{Imag}(z_i))$$$. Then, the k-space data is modeled as $$$y=\Phi z + \varepsilon$$$, where $$$\Phi=EW$$$.


Inference: We perform inference on this model using the Expectation-Maximization (EM) algorithm. We iteratively alternate between an E-Step, which solves for a posterior distribution $$$p(z|y)=\mathcal{N}(\mu, \Sigma)$$$, and an M-Step, which updates the regularization penalties $$$\alpha$$$: $$\mu=\beta\Sigma \Phi^\top y, \quad\Sigma=\left(\beta\Phi^\top \Phi + \text{diag}(\alpha))\right)^{-1}\quad\text{(E-Step)}$$ $$\alpha_i=\frac{2}{\text{Real}(\mu_i)^2+\text{Real}(\Sigma_{i,i})+\text{Imag}(\mu_i)^2+\text{Imag}(\Sigma_{i,i})} \forall i=1,2,\ldots,N\quad\text{(M-Step)}$$ For images with a large number of pixels $$$N$$$, directly applying these EM updates can be very expensive due to the matrix inversion needed by the E-Step. To resolve this issue, we follow [6, 7] in using the conjugate gradient (CG) algorithm to directly solve for $$$\mu$$$ and probe the diagonal elements of $$$\Sigma$$$ without needing to compute $$$\Sigma$$$ itself. Thus, the time complexity of Bayes-SENSE inference scales as $$$O(N\log N)$$$, which is asymptotically the same as L2-SENSE/L1-SENSE.

Image Reconstruction and Variance Map: After running EM for $$$T$$$ iterations, we can recover a posterior distribution for the image as $$$x|y\sim\mathcal{N}(\mu_x,\Sigma_x)$$$, where $$$\mu_x=W\mu$$$ and $$$\Sigma_x=W\Sigma W^\top$$$. The mode of this distribution $$$\mu_x$$$ is our reconstructed image and the diagonal elements of $$$\Sigma_x$$$ form a pixel-wise variance map.

Unlike L2-SENSE and L1-SENSE, Bayes-SENSE does not require tuning of regularization penalties because its penalties $$$\alpha$$$ are learned from the data. Like BCS [5, 6], Bayes-SENSE is easy to extend to multi-contrast reconstruction by having all contrasts share a common $$$\alpha$$$ vector of regularization penalties. We call this extension Joint Bayes-SENSE, which can efficiently parallelize the multi-contrast E-Steps across multi-core machines.

Experiments

We compare L2-SENSE, L1-SENSE, and Bayes-SENSE on several reconstruction tasks. In all cases, we report best results after tuning the regularization penalty $$$\lambda$$$ for L2-SENSE and for L1-SENSE. For L1-SENSE and Bayes-SENSE, we use 3-level Daubechies-2 wavelets. We run Bayes-SENSE for $$$T=8$$$ iterations with $$$K=10$$$ CG probe vectors. All acquisitions are made with 32ch coils. Code/data can be found at https://anonymous.4open.science/r/bayes-sense-BF92/. We use the SigPy package [8] for L2-/L1-SENSE implementations.

Fig2. presents results for reconstruction of (1) a 256x256 in-vivo FLAIR image with uniform undersampling in 1D at $$$R=4$$$ and (2) a 224x210 in-vivo $$$T_2*$$$-weighted image with Poisson undersampling in 2D at $$$R=8$$$. In both cases, we show that Bayes-SENSE outperforms L2- and L1-SENSE in NRMSE, while not needing $$$\lambda$$$-tuning. Bayes-SENSE also generates variance maps that correlate well with reconstruction error.

Fig3. presents results for reconstruction of three 256x256 in-vivo images (GRE, $$$T_2$$$-weighted, FLAIR) from $$$R=5$$$ uniform undersampling in 1D at $$$R=5$$$. Joint Bayes-SENSE outperforms L2-SENSE and slightly outperforms L1-SENSE in terms of NRMSE, while automatically tuning regularization penalties for all contrasts simultaneously.

Fig4. presents results for reconstruction of four 224x210 in-vivo $$$T_2*$$$-weighted echoes with Poisson undersampling in 2D at $$$R=12$$$. Joint Bayes-SENSE outperforms both L2-SENSE (not pictured) and L1-SENSE in NRMSE on all contrasts by leveraging the shared information among the contrasts.

Fig5. We repeat the experiment in Fig4 for undersampling factors $$$R=4,6,8,10,12,14,16$$$. Joint Bayes-SENSE consistently outperforms L2-SENSE and L1-SENSE in terms of overall NRMSE. To process four contrasts, L2-SENSE/L1-SENSE take ~2-3 minutes on CPU while Bayes-SENSE takes ~3-4 minutes on GPU.

Discussion/Conclusion

In this work, we introduced Bayesian Sensitivity Encoding, which can improve upon L2-/L1-SENSE by providing automatic regularization tuning, uncertainty quantification, and the ability to integrate multi-contrast data for joint reconstruction. Future work will tackle further reducing the computational cost of Bayes-SENSE by exploiting preconditioning for CG. As the memory footprint of the reconstruction is small, several slices can be processed in parallel on GPU, thereby providing further speed-up.

Acknowledgements

This work was supported by research grants NIH R01 EB028797, R03 EB031175, U01 EB025162, P41 EB030006, U01 EB026996, and the NVidia Corporation and Amazon AWS Research Awards for computing support.

References

[1] Pruessmann, K. P., Weiger, M., Scheidegger, M. B., & Boesiger, P. (1999). SENSE: sensitivity encoding for fast MRI. Magnetic Resonance in Medicine: An Official Journal of the International Society for Magnetic Resonance in Medicine, 42(5), 952-962.

[2] Lin, F. H., Kwong, K. K., Belliveau, J. W., & Wald, L. L. (2004). Parallel imaging reconstruction using automatic regularization. Magnetic Resonance in Medicine: An Official Journal of the International Society for Magnetic Resonance in Medicine, 51(3), 559-567.

[3] Liu, B., Sebert, F. M., Zou, Y., & Ying, L. (2008, May). SparseSENSE: randomly-sampled parallel imaging using compressed sensing. In Proceedings of the 16th Annual Meeting of ISMRM.

[4] Lustig, M., Donoho, D., & Pauly, J. M. (2007). Sparse MRI: The application of compressed sensing for rapid MR imaging. Magnetic Resonance in Medicine: An Official Journal of the International Society for Magnetic Resonance in Medicine, 58(6), 1182-1195.

[5] Bilgic, B., Goyal, V. K., & Adalsteinsson, E. (2011). Multi‐contrast reconstruction with Bayesian compressed sensing. Magnetic resonance in medicine, 66(6), 1601-1615.

[6] Lin, A., Ba, D., & Bilgic, B. (2021, May). Accelerating Bayesian compressed sensing for fast multi-contrast reconstruction. In Proceedings of the 30th Annual Meeting of ISMRM.

[7] Lin, A., Song, A. H., Bilgic, B., & Ba, D. (2021). Covariance-Free Sparse Bayesian Learning. arXiv preprint arXiv:2105.10439.

[8] Ong, F., & Lustig, M. (2019). SigPy: a python package for high performance iterative reconstruction. Proceedings of the International Society of Magnetic Resonance in Medicine, Montréal, QC, 4819. Accessed at https://sigpy.readthedocs.io/en/latest/index.html.

Figures

Diagram of Bayes-SENSE reconstruction, which inputs multi-coil undersampled k-space data and outputs an image reconstruction along with a variance map. Bayes-SENSE combines the forward models for SENSE and Bayesian compressed sensing. The algorithm alternates between an E-Step to solve for the wavelet coefficients with conjugate gradient (CG) and an M-Step to update the element-wise regularization penalties. Bayes-SENSE can be easily extended to Joint Bayes-SENSE for multi-contrast reconstruction by sharing the regularization penalties across the multiple contrasts.

(Top) Bayes-SENSE outperforms L2-SENSE and L1-SENSE (with optimal $$$\lambda$$$-tuning) on 32-coil reconstruction of an in-vivo FLAIR image with uniform undersampling in 1D at R=4. (Bottom) Bayes-SENSE outperforms L2-SENSE and L1-SENSE on 32-coil reconstruction of an in-vivo $$$T_2*$$$-weighted image with Poisson undersampling in 2D at R=8. Bayes-SENSE also provides variance maps for both cases. Pixel mean squared error is observed to correlate with the pixel variance intensity.

Joint Bayes-SENSE outperforms L2-SENSE and slightly outperforms L1-SENSE while automatically tuning regularization penalties for in-vivo multi-contrast reconstruction (GRE, $$$T_2$$$-weighted, FLAIR) from 32-coil data uniformly undersampled in 1D at R=5. A representative variance map from Bayes-SENSE for one of the contrasts is also displayed.

Joint Bayes-SENSE outperforms L2-SENSE (not shown) and L1-SENSE while automatically tuning regularization penalties for four in-vivo $$$T_2*$$$-weighted echoes from 32-coil data with Poisson undersampling in 2D at R=12. L2-SENSE obtains NRMSES of {6.88, 7.09, 8.66, 9.54} (left to right). A representative variance map from Bayes-SENSE for one of the contrasts is also displayed.

NRMSE versus acceleration factor R=4,6,8,10,12,14,16 for the multi-echo reconstruction setting of Figure 4. Joint Bayes-SENSE consistently outperforms L2-SENSE and L1-SENSE. For a fixed NRMSE (i.e. horizontal line in the figure), Joint Bayes-SENSE enables higher acceleration factor than the other two methods.

Proc. Intl. Soc. Mag. Reson. Med. 30 (2022)
3444
DOI: https://doi.org/10.58530/2022/3444