3198

ProxVNET: A proximal gradient descent-based deep learning model for dipole inversion in susceptibility mapping
Christian Kames1,2, Jonathan Doucette1,2, and Alexander Rauscher1,2,3
1Department of Physics and Astronomy, University of British Columbia, Vancouver, BC, Canada, 2UBC MRI Research Centre, University of British Columbia, Vancouver, BC, Canada, 3Department of Pediatrics, University of British Columbia, Vancouver, BC, Canada

Synopsis

A deep learning model, ProxVNET, is proposed to solve the ill-posed dipole inversion in susceptibily mapping. ProxVNET is derived from unrolled proximal gradient descent iterations wherein the proximal operator is implemented as a V-Net and is itself learned. ProxVNET is shown to outperform the U-Net-based dipole inversion deep learning model QSMnet when compared to COSMOS reconstructed susceptibility maps.

Introduction

The variational minimization problem for the ill-posed dipole inversion in susceptibility mapping is given by: $$\text{argmin}_{\chi}\frac{\mu}{2}||D\chi - \varphi||_2^2+R(\chi)$$ where $$$\chi$$$ is the magnetic susceptibility, $$$D$$$ is the dipole convolution matrix, $$$\varphi$$$ the local field, and $$$R(\cdot)$$$ a regularizer. Equation [1] can be solved using a proximal gradient descent (PGD) method: $$\chi_{k+1}=\text{prox}_{R}(\chi_k-\tau_{k}D^*(D\chi-\varphi))$$ where the proximal operator prox depends on $$$R$$$. Mardani et al. proposed to learn the proximal operator using residual networks for 2D undersampled compressed sensing image reconstruction1. We adopt this strategy of learning the proximal operator for the dipole inversion, but due to memory constraints inherent to 3D problems, we have to adjust the neural network architecture. Given a zero initial input $$$\chi_0$$$, unrolling the first two iterations of the above iterative scheme yields: $$\chi_1=\text{prox}_1(D^*\varphi)$$ $$\chi_2=\text{prox}_{2}(\chi_1-\tau_{2}D^*(D\chi-\varphi))$$ where we absorbed $$$\tau_1$$$ into the first proximal operator. Ignoring the second prox operator temporarily, we are left with $$$D^*\varphi$$$ as input into a neural network whose output is passed into a single gradient descent step. As multiple iterations are increasingly memory intensive we propose to maximize the capacity of the neural network in the first step and only use the first two iterations of the learned PGD. To do so, we choose the V-Net architecture2. We demonstrate that V-Net alone outperforms the U-Net3-based QSMnet4 model. Then, we improve the V-Net model by combining the physics of the dipole problem with the expressive power of the V-Net; the proposed ProxVNET model is trained utilizing the iteration scheme in Equation [3].

Methods

Data acquisition High-resolution 3D-GRE data was acquired from 5 healthy volunteers, with 12 multiple-echo scans at varying head orientations on a Philips Achieva 3T scanner. Images were upsampled to isotropic resolution by zero-filling in the Fourier domain. Quality assurance inspection was performed on each scan in order to check for motional and foldover artifacts. A total of 48(=8+11+11+10+8) multiple-echo scans remained, with subject #4 (10x4=40 3D images) used for testing, and the remaining 4 subjects (11x4+(8+11+8)x5=179 3D images) used for training and validation. The scan parameters for all 5 datasets are summarized in Table 1. For each volunteer, the multiple-orientation reference COSMOS was computed for each echo using all remaining scan orientations. Each echo was unwrapped using Laplacian unwrapping5; the background field was removed using v-SHARP6; NiftyReg was used with default parameters to register the scans to the neutral frame7. These COSMOS reconstructions were then used as the target susceptibility maps during training and testing.
Network architecture For the proximal operator in the first iteration we use a V-Net with pre-activation (BatchNorm + ReLu) convolutional layers with 3x3x3 filters. We further incorporate a single pre-activation residual block after the first convolutional layer (before downsampling) and in front of the very last 1x1x1 convolutional layer. A skip connection between the input and output was added for residual learning. For the second proximal operator we use a 7x7x7x1x32 convolutional layer followed by a 3x3x3 convolutional layer + ReLu, another 3x3x3 convolutional layer without activation and finally a 1x1x1 convolutional layer. Networks were trained on 64x64x64 patches. Training data was generated by randomly sampling 100 patches from each of the 179, multi-orientation, multi-echo, multi-resolution images. Adam with decoupled weight decay (AdamW)8 was used to train the network, with a constant learning rate of 3e-4 and weight decay of 1e-5. The $$$\ell_1$$$ loss function was used. No further data augmentation was performed. Batch size used was 8. The network was implemented using the Julia programming language. All networks were trained using the same training data and optimization algorithm + parameters on a NVIDIA Quadro P6000 GPU with 24GB of memory.

Results

Both V-Net and ProxVNET outperform our in-house LSMR-based dipole inversion algorithm9 as well as QSMnet, as seen Table 2. V-Net and ProxVNET perform similarly well. Figure 1 shows higher visual reconstruction quality for V-Net and ProxVNET compared to QSMnet, as well.

Discussion and Conclusion

In this work we proposed a deep learning model, ProxVNET, in order to solve the ill-posed dipole inversion in susceptibily mapping. ProxVNET is derived from unrolled proximal gradient descent iterations wherein the proximal operator is a learned V-Net. Both V-Net and ProxVNET outperform the U-Net-based dipole inversion deep learning model QSMnet when compared to the COSMOS reconstructed susceptibility maps.

Acknowledgements

Natural Sciences and Engineering Research Council of Canada, Grant/Award Number 016-05371 and the Canadian Institutes of Health Research, Grant Number RN382474-418628.

References

[1] Mardani M, Sun Q, Donoho D, et al. Neural Proximal Gradient Descent for Compressive Imaging. In: Bengio S, Wallach H, Larochelle H, et al. (eds) Advances in Neural Information Processing Systems 31. Curran Associates, Inc., pp. 9573–9583.

[2] Milletari F, Navab N, Ahmadi S-A. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In: 2016 Fourth International Conference on 3D Vision (3DV). 2016, pp. 565–571.

[3] Ronneberger O, Fischer P, Brox T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In: Navab N, Hornegger J, Wells WM, et al. (eds) Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015. Springer International Publishing, 2015, pp. 234–241.

[4] Yoon J, Gong E, Chatnuntawech I, et al. Quantitative susceptibility mapping using deep neural network: QSMnet. NeuroImage 2018; 179: 199–206.

[5] Schofield MA, Zhu Y. Fast phase unwrapping algorithm for interferometric applications. Opt Lett, OL 2003; 28: 1194–1196.

[6] Wu B, Li W, Guidon A, et al. Whole brain susceptibility mapping using compressed sensing. Magnetic Resonance in Medicine 2012; 67: 137–147.

[7] Modat M, Cash DM, Daga P, et al. Global image registration using a symmetric block-matching approach. JMI 2014; 1: 024003.

[8] Loshchilov I, Hutter F. Decoupled Weight Decay Regularization. arXiv:171105101 [cs, math], http://arxiv.org/abs/1711.05101 (2017, accessed 24 May 2019).

[9] Kames C, Wiggermann V, Rauscher A. Rapid two-step dipole inversion for susceptibility mapping with sparsity priors. NeuroImage 2018; 167: 276–283.

Figures

Table 1: Summary of dataset used for training and testing the neural network.

Table 2: Comparison of the normalized root-mean-squared error (NRMSE) and peak signal-to-noise ratio (PSNR) error metrics for QSMnet, V-Net, the proposed ProxVNET, and a traditional LSMR-based method. V-Net and ProxVNET perform similarly well, outperforming the other methods.

Figure 1: Comparison of LSMR, QSMnet, ProxVNET, V-Net, and COSMOS susceptibility reconstructions.

Proc. Intl. Soc. Mag. Reson. Med. 28 (2020)
3198