3780

Q-space trajectory imaging with positivity constraints: a machine learning approach
Deneb Boito1,2 and Evren Özarslan1,2
1Biomedical Engineering, Linköping University, Linköping, Sweden, 2Center for Medical Image Science and Visualization, Linköping University, Linköping, Sweden

Synopsis

Keywords: Data Processing, Diffusion/other diffusion imaging techniques

Q-space trajectory imaging (QTI) is a diffusion MRI framework which access features of the microstructure through the statistical moments of the diffusion tensor distribution. To overcome unreliable estimates obtained with standard fitting methods, a constrained estimation framework named QTI+ was recently proposed. Constrained optimization however typically requires sophisticated fitting routines which introduce a heavy computational burden. In this work we thus explore the possibility of speeding up the QTI parameter estimation, while retaining strict positivity constraints, using artificial intelligence. Results are shown on synthetic datasets as well as for healthy subjects and data from brain tumor patients.

Introduction

Q-space trajectory imaging (QTI)1 is a diffusion MRI (dMRI) framework combining tensor-encoded diffusion measurements and a diffusion tensor distribution (DTD)2 model for the tissue microstructure. The MR signal is represented via cumulant expansion as:
$$ln(S)\,\approx\,ln(S_0)\,-\mathbf{B}:\mathbf{\hat{D}}\,+\frac{1}{2}\,\mathbb{B}:\mathbb{C},$$ where $$$S$$$ is the acquired signal, $$$S_0$$$ is the signal for null diffusion gradients, $$$\mathbf{\hat{D}}$$$ and $$$\mathbb{C}$$$ are, respectively, the second order mean and fourth order covariance tensors of the DTD, and $$$\mathbf{B}$$$ are the measurement tensors. A constrained estimation framework, called QTI+, which imposes relevant positivity conditions on the estimated QTI quantities, was recently introduced by Herberthson et al.3 . Estimates obtained with this framework were shown to both increase the method’s robustness with respect to noise, and lower the requirement on the number of measurements to be collected for the fitting3,4. However, constrained estimation typically leads to lengthier computations. As previously done by other researchers in the dMRI community5,6,7,8, in this work we harness the computational efficiency of neural networks to speed up the estimation of QTI parameters while enforcing strict positivity conditions considered in QTI+. We show that QTI metrics produced by the network are comparable to those obtained with QTI+, providing overall smoother maps via shorter computation times.

Methods

$$$\textbf{Positivity}\,\textbf{conditions}$$$
The positivity conditions considered in QTI+ were named d), c), and m) according to which tensor they are enforced upon ($$$\mathbf{\hat{D}},\mathbb{C}\,,$$$ and $$$\,\mathbb{M}=\mathbf{\hat{D}}\otimes\mathbf{\hat{D}}\,+\,\mathbb{C}$$$, respectively). These conditions can be enforced in several steps involving Semidefinite Programming (SDP) and non-linear least squares (NLLS) routines. In this work we consider QTI+ results obtained with the NLLS(dc) routine, which enforces conditions d) and c) using NLLS with initial guess provided by an SDP routine.

$$$\textbf{Neural}\,\textbf{Network}\,\textbf{implementation}$$$
A neural network was implemented using Keras9 in TensorFlow10 in an encoder/decoder architecture. The encoder is a MultiLayer Perceptron (MLP) whose final output is 28 numbers interpreted as being $$$S_0\,,L\,$$$and$$$\,\Gamma$$$, where $$$L$$$ and $$$\Gamma$$$ are lower triangular matrices representing the Cholesky Factorization of $$$\mathbf{\hat{D}}$$$ and $$$\mathbb{C}$$$. The $$$\mathbf{\hat{D}}\,$$$and$$$\,\mathbb{C}$$$ tensors are then obtained as $$$\mathbf{\hat{D}}\,=\,LL^\text{T}\,$$$and$$$\,\mathbb{C}\,=\,\Gamma\Gamma^\text{T}$$$, thus ensuring the satisfaction of conditions d) and c) from QTI+. The decoder then reconstructs the predicted dMRI signal from $$$S_0,\,\mathbf{\hat{D}}$$$, and $$$\mathbb{C}$$$. The network’s architecture, hereafter referred to as ML(dc), and the hyperparameters are displayed in Figure(1). Note that such network is meant to be trained directly on the data of interest, and as such its performance is not affected by possible biases present in the training data. However, it is still possible to train the network on either synthetic or real data, and subsequently deploy it on unseen data to gain additional computational speed.

$$$\textbf{Simulations}\,\textbf{on}\,\textbf{realistic}\,\textbf{brain}\,\textbf{data}$$$
Two realistic healthy brain datasets were generated by computing the QTI signal from parameters obtained by employing QTI+ on a publicly available datasets11 and its subsampled versions. Hereafter these are referred to as p217 and p56 according to the number of diffusion measurements included in the respective sets. Noise from the Rician distribution was added to the synthetic data, hereafter referred to as p217synth and p56synth, to obtain a Signal-to-Noise ratio of 25. The NLLS(dc) and ML(dc) routines’ performances were assessed by computing the difference between the ground truth QTI maps and those obtained with the two method.

$$$\textbf{Clinical}\,\textbf{and}\,\textbf{experimental}\,\textbf{data}$$$
Data from two patients referred for tumor surgery were included in the study after written informed consent (Ethics approval: EPM$$$\,2020-01404$$$). Two protocols were used for the acquisition of the dMRI data: one tumor dataset consisted of 104 diffusion volumes with voxel size $$$2.4\times2.4\times4.8\,\,mm^3$$$, while the other dataset consisted of 70 diffusion volumes with voxel size $$$2.2\times2.2\times2.4\,\,mm^3$$$. Data from a healthy volunteer were also collected with both protocols (Ethics approval:$$$\,2018/143-32$$$). All datasets were preprocessed for motion, eddy currents, and EPI-distortions using tools from FSL12.

Results

The results in Figure(2) show that ML(dc) produces estimates of similar quality to those of NLLS(dc) in terms of derived QTI-metrics for p217synth, while being markedly better on p56synth. The results also show that ML(dc) exhibits a bias towards reducing anisotropy (quantified by Fractional Anisotropy$$$\,(\textrm{FA})$$$, Microscopic Fractional Anisotropy$$$\,(\mu\textrm{FA})$$$, and Orientational Coherence ($$$\textrm{C}_\text{c}$$$) and increasing Mean Diffusivity ($$$\textrm{MD}$$$) and Size Variance ($$$\textrm{C}_\text{MD}$$$). This is probably due to an intrinsic denoising process occurring during network training.

The maps in Figure(3) show superior quality for the estimates produced with ML(dc), especially when fewer dMRI volumes are available.

Figures(4)$$$\,$$$and$$$\,$$$(5) show the results obtained on the two tumor datasets, for both a network trained directly on the tumor data (ML(dc)), and a network trained on the healthy volunteer data and then deployed on the patients’ data (ML*(dc)). Despite higher fitting residuals, the ML*(dc) net produces similar QTI maps to ML(dc). Compared to NLLS(dc), both ML strategies produce visibly smoother results which seems to better highlight certain details about the tumor’s microenvironments, but at the same time could be removing precious details. Computations took on average 0.4 seconds for ML*(dc), 3 minutes for ML(dc), and 40 minutes for NLLS(dc).

Discussion & Conclusion

Similarly to previous dMRI studies5,6,7,8, this work also suggests neural networks as a valuable alternative to standard fitting routines, especially when computational time is a concern. In conclusion, we presented a neural network architecture in which model constraints are enforced. This approach could be applied to other dMRI methods for which positivity constraints are available13.

Acknowledgements

The authors thank the staff at the Department of Neurosurgery at Linköping University Hospital and Elisabeth Klint, MSc, and Karin Wårdell, PhD, from the Department of Biomedical Engineering, Linköping University, for data collection. This project was financially supported by the Swedish Foundation for Strategic Research (RMX18-0056), Linköping University Center for Industrial Information Technology (CENIIT), Sweden's Innovation Agency (VINNOVA) ASSIST, and Analytic Imaging Diagnostic Arena (AIDA).

References

1. Westin, C.F., Knutsson, H., Pasternak, O., Szczepankiewicz, F., Özarslan, E., van Westen, D., Mattisson, C., Bogren, M., O'Donnell, L.J., Kubicki, M., Topgaard, D., Nilsson, M.. Q-space trajectory imaging for multidimensional diffusion MRI oft he human brain. NeuroImage 2016;135:345

2. Jian, B., Vemuri, B.C., Özarslan, E., Carney, P.R., Mareci, T.H.. A novel tensor distribution model for the diffusion-weighted MR signal. NeuroImage 2007;37(1):164-176.

3. Herberthson M., Boito D., Haije TD, Feragen A., Westin CF, Özarslan E. Q-space trajectory imaging with positivity constraints (QTI+). Neuroimage. 2021 Sep;238:118198. doi: 10.1016/j.neuroimage.2021.118198. Epub 2021 May 21. PMID: 34029738.

4. Boito D., Herberthson M., Dela Haije T. Özarslan E. Enforcing positivity constraints in q-space trajectory imaging (QTI) allows for reduced scan time. In Proc Intl Soc Mag Reson Med, volume 29, page 0404, 2021.

5. Barbieri S., Gurney-Champion OJ, Klaassen R., Thoeny HC. Deep learning how to fit an intravoxel incoherent motion model to diffusion-weighted MRI. Magn Reson Med. 2020 Jan;83(1):312-321. doi: 10.1002/mrm.27910. Epub 2019 Aug 7. PMID: 31389081.

6. Reisert M., Kellner E., Dhital B., Hennig J., Kiselev V.G.. Disentangling micro from mesostructure by diffusion MRI: a Bayesian approach. Neuroimage, 147 (2017), pp. 964-975

7. Palombo M., Ianus A., Guerreri M., Nunes D., Alexander D.C., Shemesh N., Zhang H.. SANDI: a compartment-based model for non-invasive apparent soma and neurite imaging by diffusion MRI. Neuroimage, 215 (2020), Article 116835

8. de Almeida Martins J.P. , Nilsson M., Lampinen B., Palombo M., While P.T., Westin C.F., Szczepankiewicz F.. Neural networks for parameter estimation in microstructural mri: application to a diffusion-relaxation model of white matter. Neuroimage, 244 (2021), p. 118601

9. Chollet, F., et al. (2015). Keras. GitHub. Retrieved from https://github.com/fchollet/keras

10. Abadi M. et al., TensorFlow: Large-scale machine learning on heterogeneous systems, 2015. Software available from tensorflow.org.

11. Szczepankiewicz F, Hoge S, Westin CF. Linear, planar and spherical tensor-valued diffusion MRI data by free waveform encoding in healthy brain, water, oil and liquid crystals. Data Brief. 2019;25:104208.

12. Jesper L. R. Andersson and Stamatios N. Sotiropoulos. An integrated approach to correction for off-resonance effects and subject movement in diffusion MR imaging. NeuroImage, 125:1063-1078, 2016.

13. Dela Haije, T., Özarslan, E., Feragen, A. Enforcing necessary nonnegativity constraints for common diffusion MRI models using sum of squares programming. NeuroImage 2020;209:116405.

Figures

The neural network’s architecture and the hyperparameters used for training. A MLP with three hidden layers compresses the input signal into the 28 QTI parameters (1 for $$$S_0$$$, 6 for $$$\mathbf{\hat{D}} = LL^\text{T}$$$, 21 for $$$\mathbb{C} = \Gamma\Gamma^\text{T}$$$), from which the predicted dMRI signal is reconstructed to compute the loss. Batch size and number of epochs were set to avoid lengthy training times and over smoothing of the maps derived from the estimated QTI parameters.

Results from the simulations on realistic brain data. Differences are computed by subtracting each estimated metric from its ground truth value. ML(dc) provides results comparable to NLLS(dc) for p217synth, while outperforming it on p56synth. On both datasets the fits produced with ML(dc) show a tendency towards underestimating the anisotropy metrics ($$$\textrm{FA}, \mu\textrm{FA}, \textrm{C}_c$$$), while overestimating $$$\textrm{MD}$$$ and $$$\textrm{C}_{\text{MD}}$$$. This is seemingly due to the denoising and smoothing effects introduced by the MLP compression.

Results on the p217 and p56 datasets. The maps obtained from the ML(dc) fits appear smoother and less noisy for both datasets. Particularly for the $$$\textrm{C}_\textrm{MD}$$$ and $$$\textrm{C}_{\textrm{c}}$$$ maps, the network produces markedly improved results, especially for the p56 dataset. The rightmost column shows that the m) condition from QTI+, which is not imposed with NLLS(dc) nor ML(dc), is violated (green colored voxels) with similar frequency and in similar locations for the two methods.

Results on one of the tumor datasets. The top row displays the standard clinical images. The middle part shows the QTI derived maps for NLLS(dc), ML(dc), and ML*(dc) fits. Fit residuals are shown in the rightmost column, while difference maps between NLLS(dc) and the two network fits are shown at the bottom. QTI maps obtained with the two networks appear smoother and denoised, which seems to better display features of the microstructure. However, this could also result in loss of information and/or reduced contrast, as it seems to happen for example in the $$$\text{C}_{\text{MD}}$$$ map.

Results on one of the tumor datasets. The top row displays the standard clinical images. The middle part shows the QTI derived maps for NLLS(dc), ML(dc), and ML*(dc) fits. Fit residuals are shown in the rightmost column, while difference maps between NLLS(dc) and the two network fits are shown at the bottom. Interestingly, $$$\mu\textrm{FA}$$$ and $$$\textrm{C}_{\textrm{MD}}$$$ maps appear drastically different between the ML and NLLS(dc) fits inside the tumor region.

Proc. Intl. Soc. Mag. Reson. Med. 31 (2023)
3780
DOI: https://doi.org/10.58530/2023/3780