5193

Constraint function for IVIM quantification using unsupervised learning
Wonil Lee1, Beomgu Kang1, Jongyeon Lee1, Georges El Fakhri2, Chao Ma2, Yeji Han3, Jun-Young Chung4, Young Noh5, and HyunWook Park1
1The School of Electrical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Yuseong-gu, Korea, Republic of, 2Massachusetts General Hospital, Boston, MA, United States, 3Department of Biomedical Engineering, Gachon University, Incheon, Korea, Republic of, 4Department of Neuroscience, College of Medicine, Gachon University, Incheon, Korea, Republic of, 5Department of Neurology, Gil Medical Center, Gachon University College of Medicine, Incheon, Korea, Republic of

Synopsis

Keywords: Machine Learning/Artificial Intelligence, Quantitative Imaging, Intravoxel incoherent motion

Recently, various methods have been proposed to quantify intravoxel incoherent motion parameters. Many studies have shown that quantification methods using deep learning can accurately estimate IVIM parameters. Unsupervised learning is useful when quantifying IVIM parameters for in-vivo data because it does not require label data. However, in some cases, loss function does not converge as iteration increases. Constraint functions can be used to solve these problems by limiting the range of estimated outputs. In this study, we investigated the effects of constraint function to limit the range of estimated output.

Introduction

The concept of IVIM was introduced to describe diffusion and perfusion in tissue using biexponential model by Le Bihan (1,2). Although various methods for quantification of IVIM have been proposed, it is still challenging to quantify IVIM parameters accurately (3). Recently, deep learning methods have been studied to accurately quantify IVIM parameters (4,5). Unsupervised learning is useful when quantifying IVIM parameters of in-vivo data because it does not require label data. However, in some cases, loss function diverges as iteration increases. Constraint function can be used to solve these problems by limiting the range of estimated output. In this study, we investigate the effects of constraint function and scaling when quantifying IVIM parameters using unsupervised learning.

Methods

Using unsupervised learning technique, a deep neural network estimates the IVIM parameters from the diffusion weighted images. The deep neural network is trained by minimizing a loss function between the measured diffusion weighted MR signals and the synthesized diffusion weighted MR signals from the estimated IVIM parameters. IVIM model is delineated as follows.
$$S(b)=S_0( f \cdot exp(-bD_p)+(1-f)\cdot exp(-bD))$$
where $$$S(b)$$$ is the diffusion-weighted MRI signal for a b-value of $$$b$$$, $$$S(b)$$$ is the MRI signal when the b-value is zero,$$$f$$$ is the perfusion fraction, $$$D_p$$$ is the perfusion coefficient, and $$$D$$$ is the diffusion coefficient.
Various approaches have been proposed to quantify IVIM parameters by using unsupervised learning methods (4,6). In this study, IVIM parameters were quantified by Qnet, which estimated direction-dependent parameters in each orthogonal direction (6). The perfusion coefficient and diffusion coefficient, which are direction-dependent parameters, are estimated from the DWIs in the corresponding direction.
As shown in Figure 1, there are two Qnets for quantification of IVIM parameters, direction independent Qnet (Qnetf) for perfusion fraction and direction dependent Qnet (QnetDp, D) for perfusion coefficient and diffusion coefficient. The direction independent Qnet estimates perfusion fraction using all diffusion weighted images of three orthogonal diffusion gradients. The direction dependent Qnet estimates perfusion coefficient and diffusion coefficient using the diffusion weighted images of the corresponding directions. The Qnet consists of three convolutional layers with one by one kernel size. Each convolution layer has 30 kernels.
The output of Qnet is constrained by various constraint function $$$g(x)$$$. To constrain the output of Qnet, we used clipped ReLU, absolute, hyperbolic tangent, sigmoid function as $$$g(x)$$$. The constrained output of $$$g(x)$$$ is scaled to the IVIM parameter ranges. We also compared the results with and without constraints and scaling. In this study, the output of g(x) is scaled to the IVIM parameter ranges of $$$f\in[0,0.5], D\in[0,3\times10^{-3}]mm^2 /s,D_p\in [5\times10^{-3},100\times 10^{-3}]mm^2/s$$$.

Digital phantom simulations

In order to analyze the accuracy of quantification, we performed digital phantom simulations. As shown in Figure 3, we simulated a digital phantom with five different values of one parameter and the other randomly distributed digital phantoms in the ranges of $$$f\in[0.05,0.45], D\in[0.3\times10^{-3},2.7\times10^{-3}]mm^2 /s,D_p\in [10\times10^{-3},45\times 10^{-3}]mm^2/s $$$. In Figure 3, to be specific, in order to analyze the quantification performance for $$$D_x$$$, diffusion weighted MR signals are synthesized based on the five different values for $$$D_x$$$ and randomly distributed $$$f, D_y, D_z, Dp_x, Dp_y,$$$ and $$$Dp_z$$$. Then, Qnet estimated IVIM parameters with various constraint functions. Results were analyzed in the same way for $$$f$$$ and $$$D_px$$$.

In-vivo experiment

Brain MRI experiments were performed on a 3 Tesla MRI scanner. In vivo experiments were conducted with approval of the institutional review board. A spin-echo EPI sequence was used for the experiments with the imaging parameters as TR = 6000ms; TE = 83ms; number of averages = 1; FOV = 240mm$$$\times$$$240mm; matrix size = 120$$$\times$$$120; slice thickness = 2mm; and bandwidth = 1544Hz/pixel. The diffusion weighted images were acquired for b-values of 0, 10, 20, 30, 40, 50, 60, 70, 80, 90, 100, 200, 400, 700, and 1000 $$$[s/mm^2]$$$ with three orthogonal diffusion gradient directions. IVIM parameters were estimated by Qnet from the obtained diffusion weighted images. In order to analyze the performance of IVIM quantification for in-vivo data, we calculated $$$L2$$$ norm between the obtained diffusion weighted signals, $$$S(b)$$$, and the synthesized MR signals from the estimated IVIM parameters, $$$\hat{S}(b)$$$ .

Results

As shown in Figure 3, when output of Qnet was not constrained, quantification performance was very low. Furthermore, clipped ReLU function tended to saturate the output. The clipped ReLU is not suitable for finding accurate parameters. For in-vivo data, the results were different according to the constraint function. Similar to the results of digital phantom simulation, quantification was not accurate when the output of Qnet was not constrained and the clipped ReLU was used as the constraint function. On the other hand, IVIM parameters were well estimated when absolute, hyperbolic tangent, and sigmoid functions were used for constraint function. However, the prefusion fraction was estimated slightly lower when absolute was used than when hyperbolic tangent and sigmoid were used.

Discussion & Conclusion

The phantom simulation and in-vivo results showed that the quantification using unsupervised learning had high performance only when the output was constrained and scaled. Absolute function performed somewhat poorly compared to Sigmoid and Tanh. The importance of the constraint function was shown in this study. It should be used carefully when quantifying IVIM parameters using unsupervised learning.

Acknowledgements

This research was supported by a grant from the Korea Health Technology R&D Project through the Korea Health Industry Development Institute (KHIDI), funded by the Ministry of Health and Welfare, Republic of Korea (grant number: HI14C1135). This work was also supported by the Korea Medical Device Development Fund grant funded by the Korea government (the Ministry of Science and ICT; the Ministry of Trade, Industry and Energy; the Ministry of Health and Welfare; the Ministry of Food and Drug Safety) (project number: 1711138003, KMDF-RnD KMDF_PR_20200901_0041-2021-02 ).

References

1. Le Bihan D. What can we see with IVIM MRI? Neuroimage 2019;187:56-67.

2. Le Bihan D, Breton E, Lallemand D, Aubin M, Vignaud J, Laval-Jeantet M. Separation of diffusion and perfusion in intravoxel incoherent motion MR imaging. Radiology 1988;168(2):497-505.

3. Cho GY, Moy L, Zhang JL, Baete S, Lattanzi R, Moccaldi M, Babb JS, Kim S, Sodickson DK, Sigmund EE. Comparison of fitting methods and b‐value sampling strategies for intravoxel incoherent motion in breast cancer. Magnetic resonance in medicine 2015;74(4):1077-1085.

4. Barbieri S, Gurney‐Champion OJ, Klaassen R, Thoeny HC. Deep learning how to fit an intravoxel incoherent motion model to diffusion‐weighted MRI. Magnetic resonance in medicine 2020;83(1):312-321. 5. Lee W, Kim B, Park H. Quantification of intravoxel incoherent motion with optimized b‐values using deep neural network. Magnetic Resonance in Medicine 2021;86(1):230-244.

6. Lee W, Choi G, Lee J, Park H. Registration and quantification network (RQnet) for IVIM‐DKI analysis in MRI. Magnetic Resonance in Medicine 2022.

Figures

Figure 1. Schematic diagram of IVIM quantification using unsupervised learning with constrained output of quantification network (Qnet).

Figure 2. Constraint function to limit the range of the ouput

Figure 3. (a) Ground truth of IVIM parameters and estimated IVIM parameters by Qnet with various constraint functions to limit the range of output.

Figure 4. (a) IVIM quantification results of in-vivo data. (b) L2 norm between obtained diffusion weighted MR signals and synthesized diffusion weighted MR signals by estimated IVIM parameters.

Proc. Intl. Soc. Mag. Reson. Med. 31 (2023)
5193
DOI: https://doi.org/10.58530/2023/5193