1772

Learn to Better Regularize in Constrained Reconstruction
Yue Guan1, Yudu Li2,3, Xi Peng4, Yao Li1, Yiping P. Du1, and Zhi-Pei Liang2
1Institute for Medical Imaging Technology, School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, China, 2Department of Electrical and Computer Engineering, University of Illinois at Urbana-Champaign, Urbana, IL, United States, 3Beckman Institute for Advanced Science and Technology, University of Illinois at Urbana-Champaign, Urbana, IL, United States, 4Mayo Clinic, Rochester, MN, United States

Synopsis

Selecting good regularization parameters is essential for constrained reconstruction to produce high-quality images. Current constrained reconstruction methods either use empirical values for regularization parameters or apply some computationally expensive test, such as L-curve or cross-validation, to select those parameters. This paper presents a novel learning-based method for determination of optimal regularization parameters. The proposed method can not only determine the regularization parameters efficiently but also yield more optimal values in terms of reconstruction quality. The method has been evaluated using experimental data in three constrained reconstruction scenarios, producing excellent reconstruction results using the selected regularization parameters.

Introduction

Image reconstruction using a priori constraints, such as sparsity, low-rankness, and machine learning priors has become popular in recent years.1-3 The underlying regularization problem requires selection of good regularization parameters to produce high-quality images. Conventional methods often either use empirical values or apply statistical testing such as L-curve,4-5 cross-validation6-7 and SURE,8-9 to determine those parameters, which is computationally expensive. In this work, we propose a novel deep learning-based approach to solve this fundamental problem associated with constrained reconstruction. We use neural networks (NN) to learn such manifolds from training data and apply them to determine the optimal regularization parameters for newly acquired images in a computationally efficient way. Experimental studies have been performed to evaluate the proposed method.

Methods

Most constrained reconstruction problems in MRI can be formulated as:
$$\hspace{10em}\widehat{\rho}=\arg \min_{\rho} ||d-E\rho||_2^2+\sum_{i=1}^n\lambda_iR_i(\rho)\hspace{10em}(1)$$
where $$$\rho$$$ denotes the desired image function, $$$d$$$ measured data, $$$\small{E}$$$ imaging operator, $$$\small{R_{i}(\cdot)}$$$ regularization functions, and $$$\small{\lambda_{i}}$$$ regularization parameters. The paper addresses the problem of optimal selection of $$$\small{\left\{\lambda_{i}\right\}}$$$ using a learning (instead of statistical testing) based method.

The proposed method exploits the fact that given $$$\small{E}$$$, $$$\small{R_{i}(\cdot)}$$$ and a class of images, most of the image quality metrics (e.g., L-curve), as a function of $$$\small{\left\{\lambda_{i}\right\}}$$$ and $$$\rho$$$, form low-dimensional manifolds $$$\small{M(\left\{\lambda_{i}\right\}, \rho)}$$$. These low-dimensional manifolds, as illustrated in Fig. 1, can be learnt and then used for prediction for a new data set. With these predicted metrics, the optimal values for $$$\small{\left\{\lambda_{i}\right\}}$$$ are then determined by NN trained using ground truth data. In our current implementation, two image quality manifolds were learnt from training data: a) SSIM (structural similarity index measure), $$$\small{M_{\text{SSIM}}(\left\{\lambda_{i}\right\}, \rho)}$$$, and b) L-curve, $$$\small{M_{\text{Lcurve}}(\left\{\lambda_{i}\right\}, \rho)}$$$, that measures the tradeoff between data fidelity and regularization.

For a given $$$\rho$$$, if we know the solutions, $$$\small{\widehat{\rho}_{m}}$$$, to Eq.(1) for a set of values of the regularization parameters, say, $$$\small{\left\{\lambda_{i}(m)\right\}_1^M}$$$, within their permissible range, $$$\small{M_{\text{Lcurve}}(\left\{\lambda_{i}(m)\right\}, \widehat{\rho}_{m})}$$$ can be calculated easily; $$$\small{M_{\text{SSIM}}(\left\{\lambda_{i}(m)\right\}, \widehat{\rho}_{m})}$$$ can also be determined if a reference $$$\small{{\rho}_{\text{ref}}}$$$ is available. However, solving Eq.(1) for $$$\small{\left\{\lambda_{i}(m)\right\}_1^M}$$$ to determine $$$\small{\widehat{\rho}_{m}}$$$ is computationally expensive and thus not feasible practically.

We solved this problem using a machine learning-based method. More specifically, we solved Eq. (1) for only a few values from $$$\small{\left\{\lambda_{i}(m)\right\}_1^M}$$$ , say $$$n+1$$$ values with $$$n$$$ being the number of regularization terms in Eq. (1). With these reconstructions, we used an interpolation method to generate $$$\small{\widehat{\rho}_{m}^0}$$$. Interestingly, while $$$\small{\widehat{\rho}_{m}^0}$$$ are rather poor approximations of $$$\small{\widehat{\rho}_{m}}$$$, we can train a neural network to map $$$\small{M_{\text{Lcurve}}(\left\{\lambda_{i}(m)\right\},\widehat{\rho}_{m}^0)}$$$ to $$$\small{M_{\text{Lcurve}}(\left\{\lambda_{i}(m)\right\}, \widehat{\rho}_{m})}$$$. Similarly, we can also train a network to map $$$\small{M_{\text{SSIM}}(\left\{\lambda_{i}(m)\right\}, \widehat{\rho}_{m}^0)}$$$ to $$$\small{M_{\text{SSIM}}(\left\{\lambda_{i}(m)\right\}, \widehat{\rho}_{m})}$$$ using references generated from $$$\small{\widehat{\rho}_{m}^0}$$$ by another network. Here, we used a generative adversarial network to produce the references, which was trained with ground truth data; we used two fully-connected networks to predict SSIM and L-curve, which were trained using training data with pre-calculated quality metrics for both $$$\small{\widehat{\rho}_{m}}$$$ and $$$\small{\widehat{\rho}_{m}^0}$$$. With SSIM and L-curve predicted, we fused them to produce the final estimate of $$$\small{\left\{\lambda_{i}\right\}}$$$ using a fully-connected network.

It is important to note that instead of directly learning the manifolds from the initial reconstructions $$$\small{\widehat{\rho}_{m}^0}$$$ or learning $$$\small{\widehat{\rho}_{m}}$$$ from $$$\small{\widehat{\rho}_{m}^0}$$$, we converted them to the initial quality metrics $$$\small{M_{\text{Lcurve}}(\left\{\lambda_{i}(m)\right\},\widehat{\rho}_{m}^0)}$$$ and $$$\small{M_{\text{SSIM}}(\left\{\lambda_{i}(m)\right\},\widehat{\rho}_{m}^0)}$$$ , thus significantly reducing the learning complexity and enhancing the quality of the predicted results of the trained networks. Also, as compared to the traditional L-curve method, the proposed method incorporates both $$$\small{M_{\text{Lcurve}}}$$$ and $$$\small{M_{\text{SSIM}}}$$$, thus enabling more optimal parameter selection.

Results

The proposed method was tested in three reconstruction scenarios: image deblurring, parallel imaging, and dynamic image denoising. The first two problems entailed one regularization term, absorbing spatial smoothness constraint; the third application involved two regularization terms imposing low-rankness and sparsity constraints, respectively.

For image deblurring, the measured data were simulated using T1W images from HCP database with a Gaussian smoothing kernel plus random noise. Figure 2 shows the reconstruction results using the regularization parameters selected by L-curve and the proposed method. As can be seen, our method led to much better reconstruction quality in much shorter processing time (L-Curve: 3s, Proposed: 0.1s per slice).

For parallel imaging, we used the multi-coil knee data from the NYU database with retrospective undersampling (R=2.5). The results are summarized in Fig. 3. As can be seen, our method achieved superior performance over the traditional L-curve method.

For dynamic image denoising, a series of $$$\small{\text{T}_2^*\text{W}}$$$ images were acquired using the mGRE sequence with 74 echoes. Gaussian noise was also added retrospectively to produce a range of SNR levels. The reconstruction results are shown in Fig. 4. Note our method handled the reconstruction problem well even with multiple regularization parameters that are very challenging to be determined optimally and efficiently using traditional statistical testing-based methods.

Conclusion

We have proposed a novel learning-based method for efficient selection of optimal regularization parameters for constrained image reconstruction. Our method has been evaluated using experimental data, producing very encouraging results. The proposed method may help improve the effectiveness and practical utility of constrained reconstruction.

Acknowledgements

This work was supported in a part by National Natural Science Foundation of China (62001293)

References

1. Meng Z, Guo R, Li Y, et al. Accelerating T2 mapping of the brain by integrating deep learning priors with low‐rank and sparse modeling. Magn Reson Med. 2020, 85(3):1455-1467.

2. Lam F, Li Y, Guo R, et al. Ultrafast magnetic resonance spectroscopic imaging using SPICE with learned subspaces. Magn Reson Med. 2020, 83(2): 377-390.

3. Wang S, Su Z, Ying L, et al. Accelerating magnetic resonance imaging via deep learning. 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI). IEEE, 2016: 514-517.

4. Hansen PC, O’Leary DP. The use of the L-curve in the regularization of discrete ILL-posed problems. SIAM J Sci Comput. 1993;14:1487–1503.

5. Vogel CR. Non-convergence of the L-curve regularization parameter selection method. Inverse Problems. 1996;12:535–547

6. Carew JD, Wahba G, Xie X, et al. Optimal spline smoothing of fMRI time series by generalized crossvalidation. Neuroimage. 2003;18:950–961.

7. Sourbron S, Luypaert R, Schuerbeek PV, et al. Choice of the regularization parameter for perfusion quantification with MRI. Phys Med Biol. 2004;49:3307–3324.

8. Ramani S, Liu Z, Rosen J, et al. Regularization parameter selection for nonlinear iterative image restoration and MRI reconstruction using GCV and SURE-based methods. IEEE T Image Process. 2012;21:3659-3672.

9. Weller DS, Ramani S, Nielsen J-F, Fessler JA. Monte carlo SURE-based parameter selection for parallel magnetic resonance imaging reconstruction. Magn Reson Med. 2014;71:1760–1770.

Figures

Figure 1. An illustration of image quality manifolds for (a) SSIM, (b) data fidelity, and (c) norm of regularization function. As can be seen, all image quality metrics, as a function of λ and ρ, reside in a low-dimensional manifold and thus learnable from training data.

Figure 2. Comparison of reconstruction results obtained from L-curve and the proposed method for image deblurring. Note the reconstruction error was significantly reduced by the proposed method, demonstrating its effectiveness in learning the optimal regularization parameters.

Figure 3. Comparison of reconstruction results from L-Curve and the proposed method for parallel imaging with under-sampling by a factor of 2.5. As can be seen, our method handled the practical reconstruction problem with superior performance.

Figure 4. Comparison of reconstructions for dynamic image denoising that includes two regularization terms imposing low-rankness and sparsity, respectively. As can be seen, our method handled the reconstruction problem well with multiple regularization parameters that are very challenging to determine optimally and efficiently using traditional statistical testing-based methods.

Proc. Intl. Soc. Mag. Reson. Med. 29 (2021)
1772