Keywords: Machine Learning/Artificial Intelligence, Machine Learning/Artificial Intelligence
Synthesis of missing contrasts in an MRI protocol via translation from acquired contrasts can reduce costs associated with prolonged exams. Current learning-based translation methods are predominantly based on generative adversarial networks (GAN) that implicitly characterize the distribution of the target contrast, with limits fidelity of synthesized images. Here we present SynDiff, a novel conditional adversarial diffusion model for computationally efficient, high-fidelity contrast translation. SynDiff enables training on unpaired datasets, thanks to its cycle-consistent architecture with coupled diffusion processes. Demonstrations on multi-contrast MRI datasets indicate the superiority of SynDiff against competing GAN and diffusion models.1. B. Moraal et al., “Multi-contrast, isotropic, single-slab 3d MR imaging in multiple sclerosis,” Eur. Radiol., vol. 18, pp. 2311–2320, 2008.
2. J. E. Iglesias, E. Konukoglu, D. Zikic, B. Glocker, K. Van Leemput, and B. Fischl, “Is synthesizing MRI contrast useful for inter-modality analysis?,” in Med. Image Comput. Comput. Assist.Interv., pp. 631–638, 2013.
3. A. Jog, A. Carass, S. Roy, D. L. Pham, and J. L. Prince, “Random forest regression for magnetic resonance image synthesis,” vol. 35, pp. 475–488, 2017.
4. T. Joyce, A. Chartsias, and S. A. Tsaftaris, “Robust multi-modal MR image synthesis,” in Med.Image Comput. Comput. Assist. Interv., pp. 347–355, 2017.
5. J. Denck, J. Guehring, A. Maier, and E. Rothgang, “Enhanced magnetic resonance image synthesis with contrast-aware generative adversarial networks,” Journal of Imaging, vol. 7, p. 133, 08 2021.
6. Wang G, Gong E, Banerjee S, Martin D, Tong E, Choi J, Chen H, Wintermark M, Pauly JM, Zaharchuk G. Synthesize High-Quality Multi-Contrast Magnetic Resonance Imaging From Multi-Echo Acquisition Using Multi-Task Deep Generative Model. IEEE Trans Med Imaging. 2020 Oct;39(10):3089-3099.
7. Lee, D., Moon, WJ. & Ye, J.C. Assessing the importance of magnetic resonance contrasts using collaborative generative adversarial networks. Nat Mach Intell 2, 34–42 (2020).
8. Kim S, Jang H, Hong S, Hong YS, Bae WC, Kim S, Hwang D. Fat-saturated image generation from multi-contrast MRIs using generative adversarial networks with Bloch equation-based autoencoder regularization. Med Image Anal. 2021 Oct;73:102198.
9. S. U. Dar, M. Yurt, L. Karacan, A. Erdem, E. Erdem, and T. Çukur, “Image synthesis in multi-contrast MRI with conditional generative adversarial networks,” IEEE Transactions on Medical Imaging, vol. 38, no. 10, pp. 2375–2388, 2019.
10. B. Yu, L. Zhou, L. Wang, J. Fripp, and P. Bourgeat, “3D cGAN based cross-modality MR image synthesis for brain tumor segmentation,” Int. Symp. Biomed. Imaging, pp. 626–630, 2018.
11. H. Li, J. C. Paetzold, A. Sekuboyina, F. Kofler, J. Zhang, J. S. Kirschke, B. Wiestler, and B. Menze, “DiamondGAN: Unified multi-modal generative adversarial networks for MRI sequences synthesis,” in Med. Image Comput. Comput. Assist. Interv., pp. 795–803, 2019.
12. O. Dalmaz, M. Yurt, and T. Çukur, "ResViT: Residual Vision Transformers for Multimodal Medical Image Synthesis," in IEEE Transactions on Medical Imaging, vol. 41, no. 10, pp. 2598-2614, Oct. 2022.
13.Pan, Y., Liu, M., Lian, C., Zhou, T., Xia, Y., and Shen, D. (2018). Synthesizing Missing PET from MRI with Cycle-consistent Generative Adversarial Networks for Alzheimer's Disease Diagnosis. MICCAI International Conference on Medical Image Computing and Computer-Assisted Intervention, 11072, 455-463
14.Chartsias, A., Joyce, T., Dharmakumar, R., Tsaftaris, S.A., 2017. Adversarial image synthesis for unpaired multi-modal cardiac data, in: Simulation and Synthesis in Medical Imaging. pp. 3–13.
15.Hiasa, Y. et al. 2018. Cross-Modality Image Synthesis from Unpaired Data Using CycleGAN, Simulation and Synthesis in Medical Imaging. pp. 31–41
16. P. Dhariwal and A. Nichol, “Diffusion models beat gans on image synthesis,” in Advances in neural Information Processing Systems, vol. 34, pp. 8780–8794, Curran Associates, Inc., 2021.
17. J. Ho, A. Jain, and P. Abbeel, “Denoising diffusion probabilistic models,” in Advances in neural information Processing Systems, vol. 33, pp. 6840–6851, 2020.
18. B. H. Menze et al., "The Multimodal Brain Tumor Image Segmentation Benchmark (BRATS)," in IEEE Trans. Med. Imaging, vol. 34, no. 10, pp. 1993- 2024, Oct. 2015.
19. O. Oktay, J. Schlemper, L. L. Folgoc, M. J. Lee, M. Heinrich, K. Misawa, K. Mori, S. G.McDonagh, N. Hammerla, B. Kainz, B. Glocker, and D. Rueckert, “Attention U-Net: Learning where to look for the pancreas,” arXiv:1804.03999, 2018.
20. H. Zhang, I. Goodfellow, D. Metaxas, and A. Odena, “Self-attention generative adversarial networks,” in Int. Conf. Mach. Learn., vol. 97, pp. 7354–7363, 2019.
21. M.-Y. Liu, T. Breuel, and J. Kautz, “Unsupervised image-to-image translation networks,” advances in Neural Information Processing Systems, vol. 30, Curran Associates, Inc., 2017.
22. X. Huang, M.-Y. Liu, S. Belongie, and J. Kautz, “Multimodal unsupervised image-to-image translation,” in Proceedings of the European conference on computer vision (ECCV), pp. 172–189, 2018.6
a) Regular diffusion models gradually transform between actual image samples and isotropic Gaussian noise in $$$T$$$ steps where $$$T$$$ is on the order of thousands. b) The proposed adversarial diffusion model instead performs fast transformation in $$$T/k$$$ steps, with step size $$$k\gg 1$$$. To capture complex reverse transition probabilities, a novel adversarial projector is introduced that leverages a diffusive generator and discriminator to synthesize a denoised image ($$$\hat{x}_{t-k}$$$), given $$$x_t$$$ and conditioned on the source contrast $$$y$$$.
Representative translation results from SynDiff and competing methods. Source, reference target, and synthesized target images are displayed for a) $$$T_2\rightarrow PD$$$ in IXI, b) $$$T_2\rightarrow FLAIR$$$ in BRATS. SynDiff provides enhanced translation performance, especially near pathological regions, in contrast to competing methods with relatively poor tissue depictions. Overall, images generated by SynDiff have better-delineated tissue boundaries with fewer artifacts and lower noise levels.