0499

NLINV-Net: Self-Supervised End-2-End Learning for Reconstructing Undersampled Radial Cardiac Real-Time Data
Moritz Blumenthal1, Guanxiong Luo1, Martin Schilling1, Markus Haltmeier2, and Martin Uecker1,3,4
1University Medical Center Göttingen, Göttingen, Germany, 2Department of Mathematics, University of Innsbruck, Innsbruck, Austria, 3Institute of Medical Engineering, Graz University of Technology, Graz, Austria, 4DZHK (German Centre for Cardiovascular Research), Göttingen, Germany

Synopsis

In this work, we propose NLINV-Net, a neural network architecture for jointly estimating the image and coil sensitivity maps of radial cardiac real-time data. NLINV-Net is inspired by NLINV and solves the non-linear formulation of the SENSE inverse problem by unrolling the iteratively regularized Gauss-Newton method, which is improved by adding neural network based regularization terms. NLINV-Net is trained in a self-supervised fashion, which is crucial for cardiac real-time data which lack any ground truth reference. NLINV-Net significantly reduces noise and streaking artifacts compared to reconstructions using plain NLINV.

Introduction

Deep neural networks were shown to improve reconstruction quality of undersampled k-space data. Physics-based neural networks1,2 reduce the amount of required training data but still require a fully sampled ground truth reference for training, which is not available for cardiac real-time data.
Recently, self-supervised reconstruction networks have been proposed which are trained without any ground truth reference by splitting the acquired k-space data in two sets and learn a reconstruction of the first set which is consistent with the second. Self-supervised networks are reported to achieve similar reconstruction quality compared to networks trained with a ground truth reference.3
In this work, we exploit the self-supervised training strategy to train NLINV-Net for reconstructing cardiac real-time data. NLINV-Net is inspired by NLINV4 and jointly estimates image and coil sensitivity maps (CSMs). The motivation to use NLINV-Net in this work is threefold:
1. Joint estimation of image and CSMs can improve classical4 and learned5 reconstruction methods.
2. NLINV has been successfully applied to cardiac real-time reconstruction6.
3. As no calibration step is required, NLINV-Net takes raw k-space data as input, making it a natural choice for the training strategy described in [3].

Theory and Methods

NLINV4 jointly estimates the image and CSMs by solving the non-linear inverse problem
$$\begin{aligned}\begin{pmatrix}\boldsymbol{x}_{\mathrm{img}}\\\boldsymbol{x}_{\mathrm{col}}\end{pmatrix}=\boldsymbol{x}=\underset{\boldsymbol{x}}{\operatorname{argmin}} \lVert F(\boldsymbol{x})-\boldsymbol{y}\rVert_2^2\,.&&&(1)\end{aligned}$$Here, $$$F$$$ is the non-linear SENSE forward operator (c.f. Fig. 1) taking the image $$$\boldsymbol{x}_{\mathrm{img}}$$$ and the CSMs $$$\boldsymbol{x}_{\mathrm{col}}$$$ as inputs and mapping them to the k-space data $$$\boldsymbol{y}$$$. The non-linear problem is solved using the iteratively regularized Gauss-Newton method (IRGNM). In each Gauss-Newton step, the iterate $$$\boldsymbol{x}^{n}$$$ is updated by solving the linearized subproblem, i.e.
$$\begin{aligned}\boldsymbol{x}^{n+1}=\underset{\boldsymbol{x}}{\operatorname{argmin}}\lVert \mathrm{D}F|_{\boldsymbol{x}^n}\left(\boldsymbol{x}-\boldsymbol{x}^n\right)-\left(\boldsymbol{y}-F\left(\boldsymbol{x}^n\right)\right)\rVert_2^2+\alpha_n\lVert\boldsymbol{x}_{\mathrm{img}}\rVert_2^2+\alpha_n\lVert W\boldsymbol{x}_{\mathrm{col}}\rVert_2^2\,.&&&(2)\end{aligned}$$ Here, $$$\mathrm{D}F|_{\boldsymbol{x}^n}$$$ is the Fréchet derivative of $$$F$$$ evaluated at $$$\boldsymbol{x}^n$$$, $$$\alpha_n$$$ is a regularization parameter reduced by a factor of two in each iteration and $$$W$$$ is a weighting matrix such that $$$\boldsymbol{x}_{\mathrm{col}}$$$ is penalized by a Sobolev norm to promote smooth CSMs.

NLINV-Net unrolls the IRGNM and includes a neural network based regularization term for $$$\boldsymbol{x}_{\mathrm{img}}$$$. The modified update rule, including a trainable regularization parameter $$$\lambda\ge 0$$$, reads $$\begin{aligned}\boldsymbol{x}^{n+1}=\underset{\boldsymbol{x}}{\operatorname{argmin}}\lVert\mathrm{D}F|_{\boldsymbol{x}^n}\left(\boldsymbol{x}-\boldsymbol{x}^n\right)-\left(\boldsymbol{y}-F\left(\boldsymbol{x}^n\right)\right)\rVert_2^2+(\alpha_n+\lambda)\lVert\boldsymbol{x}_{\mathrm{img}}-\mathrm{Net}(\boldsymbol{x}^n_{\mathrm{img}})\rVert_2^2+\alpha_n\lVert W\boldsymbol{x}_{\mathrm{col}}\rVert_2^2\,.&&&(3)\end{aligned}$$$$$\mathrm{Net}$$$ is a residual network (c.f. Fig. 2c) with 64-channel complex-valued convolutions, batch normalization and ReLU activations. Its input is normalized to a maximum magnitude of one, and the output is rescaled to the original value. The network weights are shared across iterations. The data-consistency module solving Eq. 3 (c.f. Fig. 2b) and its derivatives are implemented using the conjugate gradient algorithm. We unroll the IRGNM for $$$N=12$$$ iterations but use no network regularization (i.e. plain NLINV) in the first $$$N_{\mathrm{ini}}=N-N_{\mathrm{net}}$$$ iterations for initialization (c.f. Fig. 2a).

Self-Supervised Training Strategy
We trained NLINV-Net utilizing a real-time dataset acquired with a balanced SSFP radial FLASH sequence (TR=2.56ms, TE=1.28ms, FA=23°, FOV=256x256mm) on a Siemens Skyra 3T scanner (Siemens Healthcare GmbH, Erlangen, Germany). Multiple slices were acquired from 25 volunteers, resulting in 66.960 frames with 39 spokes each. The k-space was coil combined to $$$N_C=10$$$ virtual coils and gradient delays were estimated utilizing RING7.
Following [3], we split the acquired k-space data $$$\Omega$$$ into two disjunct subsets $$$\Theta$$$ and $$$\Lambda$$$. We experimented with two different splitting strategies, i.e. splitting the k-space spoke-wise or point-wise. Each spoke or point was selected with a probability $$$p$$$ to be contained in the first set $$$\Theta$$$, the remaining data were assigned to $$$\Lambda$$$. $$$\Theta$$$ is used by NLINV-Net to estimate the image $$$\boldsymbol{x}_{\mathrm{img}}$$$ and CSMs $$$\boldsymbol{x}_{\mathrm{col}}$$$. From that, the k-space of $$$\Lambda$$$ is predicted by the non-linear forward model with the respective pattern. The weights of NLINV-Net are optimized to minimize the normalized mean squared error of the predicted and actual k-space data in $$$\Lambda$$$ (c.f. Fig. 2.a).
We initialize the weights by training for one epoch with $$$N_{\mathrm{net}}=1$$$ network iteration. Afterwards, NLINV-Net with $$$N_{\mathrm{net}}=3$$$ or $$$N_{\mathrm{net}}=5$$$ iterations is trained for another epoch. NLINV-Net was implemented based on the new neural network framework in BART8,9. Training takes about 7, 10 and 14 hours for the $$$N_{\mathrm{net}}=1,\;3\;\text{and}\;5$$$ iterations network on an Nvida A100 GPU.

Results

We present example reconstructions using NLINV-Net with $$$N_{\mathrm{net}}=3$$$ network iterations and different training splittings in Fig. 3. As a baseline reconstruction, we provide reconstructions using plain NLINV. Since no ground truth reference is available, only a qualitative comparison of the reconstruction is possible. We observe that NLINV-Net with spoke-wise splitting reduces streaking artifact significantly compared to plain NLINV. In contrast, NLINV-Net trained with point-wise splitting is very similar to plain NLINV.
In Fig. 4, we compare reconstructions with NLINV-Net using different numbers of network iterations $$$N_{\mathrm{net}}$$$. Using $$$N_{\mathrm{net}}=3$$$ iterations slightly improves reconstruction quality compared to $$$N_{\mathrm{net}}=1$$$. The difference for $$$N_{\mathrm{net}}=3$$$ and $$$N_{\mathrm{net}}=5$$$ is minor.

Discussion and Conclusion

This work presents a NLINV-Net, a network architecture for non-linear physics-based reconstruction, and a self-supervised training strategy to train NLINV-Net without any ground truth reference data on a real-time cardiac dataset. A spoke-wise splitting strategy of the acquired k-space data seems to be superior compared to a point-wise strategy. We expect further improvement of NLINV-Net by combining NLINV-Net with - possibly learned - temporal regularization.

Acknowledgements

We acknowledge funding by the "Niedersächsisches Vorab" funding line of the Volkswagen Foundation.
This work was supported by the DZHK (German Centre for Cardiovascular Research) and funded in part by NIH under grant U24EB029240.

References

1. K. Hammernik et al., “Learning a variational network for reconstruction of accelerated MRI data”, Magnetic Resonance in Medicine, vol. 79, no. 6, pp. 3055–3071, 2018

2. H. K. Aggarwal, M. P. Mani, and M. Jacob, “MoDL: Model-Based Deep Learning Architecture for Inverse Problems”, IEEE Transactions on Medical Imaging, vol. 38, no. 2, pp. 394–405, 2019

3. B. Yaman, S. A. H. Hosseini, S. Moeller, J. Ellermann, K. Uğurbil, and M. Akçakaya, “Self-supervised learning of physics-guided reconstruction neural networks without fully sampled reference data,” Magn. Reson. Med., vol. 84, no. 6, 2020

4. M. Uecker, T. Hohage, K. T. Block, and J. Frahm, “Image reconstruction by regularized nonlinear inversion-joint estimation of coil sensitivities and image content,” Magn. Reson. Med., vol. 60, no. 3, pp. 674-682, 2008

5. M. Arvinte, S. Vishwanath, A. H. Tewfik, and J. I. Tamir, "Deep J-Sense: An unrolled network for jointly estimating the image and sensitivity maps", ISMRM Annual Meeting 2021, In Proc. Intl. Soc. Mag. Reson. Med. 29;1755, 2021

6. M. Uecker, S. Zhang, D. Voit, A. Karaus, K.-D. Merboldt, and J. Frahm, “Real-time MRI at a resolution of 20 ms,” NMR Biomed., vol. 23, no. 8, 2010, doi: 10.1002/nbm.1585.

7. S. Rosenzweig, H. C. M. Holme, and M. Uecker, “Simple auto-calibrated gradient delay estimation from few spokes using Radial Intersections (RING),” Magn. Reson. Med., vol. 81, no. 3, 2019

8. M. Uecker et al., "BART Toolbox for Computational Magnetic Resonance Imaging", Zenodo, DOI: 10.5281/zenodo.592960

9. M. Blumenthal and M. Uecker, "Deep, Deep Learning with BART", ISMRM Annual Meeting 2021, In Proc. Intl. Soc. Mag. Reson. Med. 29;1754, 2021

Figures

Fig. 1: The non-linear forward operator to be inverted by NLINV or NLINV-Net, respectively. The image $$$\boldsymbol{x}_{\mathrm{img}}$$$ and CSMs $$$\boldsymbol{x}_{\mathrm{col}}$$$ are multiplied to compute the coil images which are in turn Fourier transformed to obtain the radial k-space $$$\boldsymbol{y}$$$. In contrast to the linear formulation, the CSMs are an input of the operator rather than a parametrization.

Fig. 2: a) Network structure of NLINV-Net and self-supervised training procedure. The radial k-space is split into two subsets Θ and Λ. NLINV-Net uses Θ to jointly estimate the image and CSMs. The network parameters are optimized such that the simulated k-space matches Λ. b) Data-consistency block of NLINV-Net which computes the update due to a Gauss-Newton step utilizing the CG algorithm. c) Network structure of the denoising block consisting of five 64-channel complex-valued convolutions, batch normalization and ReLU activations acting independently on real and imaginary part.

Fig 3: Example reconstructions using 39 and 65 spokes. NLINV-Net was trained with different splitting strategies using 39 spokes per frame. It can be seen, that NLINV-Net significantly reduces noise and streaking artifacts compared to the baseline NLINV reconstruction if trained with spoke-wise splitting.

Fig 4: Example reconstructions using NLINV-Net for a varying number of network iterations $$$N_{\mathrm{net}}$$$ = 1, 3 and 5. NLINV-Net was trained using spoke-wise splitting with $$$p=0.8$$$. Using $$$N_{\mathrm{net}}=3$$$ iterations improves the reconstruction compared to $$$N_{\mathrm{net}}=1$$$, however, the difference for $$$N_{\mathrm{net}}=3$$$ and $$$N_{\mathrm{net}}=5$$$ is minor.

Proc. Intl. Soc. Mag. Reson. Med. 30 (2022)
0499
DOI: https://doi.org/10.58530/2022/0499