Matt Hemsley1,2, Brige Chugh3, Mark Ruschin3, Young Lee3, Chia-Lin Tseng4, Greg Stanisz1,2, and Angus Lau1,2
1Medical Biophysics, University of Toronto, Toronto, ON, Canada, 2Physical Sciences Platform, Sunnybrook Research Institute, Toronto, ON, Canada, 3Medical Physics, Sunnybrook Health Sciences Centre, Toronto, ON, Canada, 4Radiation Oncology, Sunnybrook Health Sciences Centre, Toronto, ON, Canada
Synopsis
The feasibility of using a neural network model to place uncertainty estimates on synthetic-CTs created with a generative adversarial network was investigated. Dropout-based variational inference was employed to account for the uncertainty on the trained model. The standard GAN loss function was also combined with an additional log-likelihood term, designed such that the network learns which regions of input data lead to highly variable output. On a dataset of n=105 brain patients, our results demonstrate that the predicted uncertainty can be interpreted as an upper bound on the true error with a confidence of approximately 95%.
Introduction
Robust electron density assignment using MR is crucial for real-time MR-only radiation treatment planning. Recently, conditional generative adversarial neural networks (cGANs) have been proposed for synthetic-CT (sCT) generation solely from MRI[1,2]. While deep neural networks are useful for prediction, existing models do not provide confidence estimates, which leads to an inability to identify incorrect outputs or out-of-distribution inputs. In the context of sCT estimation, these uncertainties can be classified into two types 1) data dependent (DD) caused by incompleteness in input data (e.g. noise, absence of visual features) and 2) model dependent (MD) due to incompleteness in the trained model (e.g. type of implant not seen in training)[3]. Uncertainty estimates would aid clinical sCT implementation by identifying failure cases where the automated output is likely incorrect and intervention (i.e. image re-acquisition or CT-based planning) is required. In this work, we investigated the feasibility of using a neural network model to place uncertainty estimates on cGAN generated synthetic-CTs.Methods
3D T1w pre-Gd, T1w post-Gd, T2 FLAIR (Matrix size 480x480, FOV=240 mm, 215-250 slices, 1mm thickness) and planning CTs (Matrix size 512x512, FOV=450 mm, 215-250 slices, 1mm thickness) of 105 brain patients were retrospectively analyzed. The MR and CT images were rigidly registered, resampled to a voxel size of (0.8x0.8x0.8)mm3, and grouped for training (n=85) and testing (n=20). The “Pix2Pix” cGAN architecture[4] was used. The model consists of two competing networks–(1) a generator (256x256 Unet) which generates candidate images based on a model distribution and (2) a discriminator (70x70 patchGAN) which discriminates between the candidate and ground truth images. The loss function is $$$L = L_{cGAN}+L_{1}$$$, where $$$L_{cGAN}$$$ captures the effectiveness of the discriminator at classifying the image, and $$$L_{1}=|Output-GroundTruth|$$$ is a data consistency term. DD uncertainty was estimated by modifying the $$$L_{1}$$$ term to include the standard deviation, $$$\sigma$$$, between the output and ground truth, $$$L_{1,Unc} = \frac{L_{1}}{\sigma}+ln(\sigma)$$$. The modified network learns both (1) the corresponding CT image, and (2) which regions of the input image lead to highly variable output. MD uncertainty was modeled using Monte Carlo dropout sampling during testing to obtain a distribution of outputs, each created with a different configuration of neurons. Testing with dropout is an approximation of Bayesian variational inference[3]. 50 sample sCTs corresponding to each MR slice with were generated, the standard deviation of the output distribution reflects the uncertainty of the model.Results
Network training took approximately 12 hours. Once trained, a sCT brain volume with DD uncertainty could be generated in 7.5 seconds, and 150 seconds for both DD and MD uncertainty. Fig1 shows the T1w MR, sCT, traditional CT, and highlights a selected region in which the network has partially failed. Fig2 shows the absolute difference map of the traditional CT and sCT, DD and MD uncertainty predictions, and indicates that areas where the network failed are regarded as uncertain. Fig3 shows the relationship between total predicted uncertainty and the true sCT-CT error in various selected regions of output. The plots demonstrate predicted uncertainty can be interpreted as an upper bound on the true error with a confidence of approximately 95%. Fig4 shows the uncertainty heatmaps highlighting regions of spatial failure in an erroneous sCT prediction.Discussion
We have demonstrated that sCT uncertainty estimates can be determined with minor adjustments to standard cGANs. MAE values obtained agree with previously published studies using similar networks, including studies on different anatomical sites[1,2]. No extra computational time was required to estimate DD uncertainty. However, MD uncertainty estimation increased computational time due to the need to generate many candidate samples. Real-time use could be enabled by parallelization with multiple GPUs. Areas of high uncertainty are predominantly edges stemming primarily from registration errors in training data, and areas where MRI offers incomplete information such as regions where bone is adjacent to air deposits, inside the bone, and the immobilization mask. As the only constraints placed on the network are training with dropout (MD) and a data regularization term in the loss function (DD) this method is not specific to cGANs and can be applied to CNNs with no adversarial component. Future work will study the use of uncertainty estimates to identify failure cases requiring the re-acquisition of images or a fall back to CT-based planning, as well as integration into dose calculations.Conclusion
This study demonstrated that sCT generation using a cGAN modified for uncertainty estimation to identify spatial regions of network failure was as robust for real-time MR-only radiation treatment planning as previous studies using traditional cGANs. The new uncertainty estimates are anticipated to enable robust sCT generation in clinical applications.Acknowledgements
The authors acknowledge funding from NSERC and Nvidia.References
1. Maspero, M., Savenije, M. H., Dinkla, A. M., Seevinck, P. R., Intven, M. P., Jurgenliemk-Schulz, I. M., . . . Cornelis A T Van Den Berg. (2018). Dose evaluation of fast synthetic-CT generation using a generative adversarial network for general pelvis MR-only radiotherapy. Physics in Medicine & Biology, 63(18), 185001. doi:10.1088/1361-6560/aada6d
2. Emami, H., Dong, M., Nejad-Davarani, S. P., & Glide-Hurst, C. K. (2018). Generating synthetic CTs from magnetic resonance images using generative adversarial networks. Medical Physics, 45(8), 3627-3636. doi:10.1002/mp.13047
3. Kendall, A., Gal, Y. (2017). What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision?. arXiv:1703.04977 [cs.CV]
4. Isola, P., Zhu, JY., Zhou T., Efros, A. (2016) Image-to-Image Translation with Conditional Adversarial Networks. arXiv:1611.07004 [cs.CV]