While gadolinium-based contrast agents (GBCAs) have been dispensable for magnetic resonance imaging in clinic practice, recently there are rising concerns about the safety of GBCAs. In previous studies, we tested the feasibility of predicting contrast agents from pre-contrast images via deep convolutional neural networks. In this study, we further improved the results with deep attention generative adversarial network. The image similarity metrics show that the model can synthesize post-contrast T1w images with superior image quality and great resemblance to the ground truth images. Restoration of contrast uptake is clearly noted on the synthesized images even in small vessels and lesions.
① Dataset and preprocessing: MR acquisitions were performed at 3T (GE Healthcare Systems, Waukesha, WI) with an 8-channel GE head coil. The MR protocol included 6 sequences (3D IR-prepped FSPGR T1w, T2w, FLAIR T2w, diffusion-weighted imaging (DWI) with 2 b-values (b=0, b=1000), T2*w and 3D ASL) acquired before injection of 0.1 mmol/kg gadobenate dimeglumine (Multihance; Bracco) and one sequence (3D IR-prepped FSPGR T1w) acquired after injection. The acquired data was co-registered to the MNI template with 2mm isotropic spatial resolution. 404 patients with different clinical indications and brain abnormalities were scanned. 323 patients chosen randomly in the cohort were used for training and the rest were used for testing. The top and bottom five slices in each case were discarded due to low signal-to-noise ratio (SNR). Figure 1 shows an example of the data that includes the 6 pre-contrast images used as input and the T1w image acquired after injection as output.
② Network architecture: The network architecture is shown in Figure 2, which follows standard conditional GAN framework with one generator and one discriminator, where the generator takes 2.5D pre-contrast images with 6 contrasts (T1w, T2w, T2*w, DWI (1), DWI (2) and ASL, 5 slices each) as input and predicts single T1w image acquired after injection. The basic component of the generator is the attention U-Net block3, which automatically learns to focus on the most important structure for contrast enhancement prediction. We designed the generator with a similar idea as “boosting” in machine learning, that a second block learns to compensate the output of the first block and they together form a more powerful generator to produce the final output, for accurate prediction of global contrast changes as well as local details. For discriminator, a 6-layer convolutional network was used.
③ Loss function: In addition to L1 loss between the synthesized image and target image, we used adversarial loss4 Ladv and perceptual loss5 Lpercep from the discriminator to encourage more realistic outputs. The overall loss function is defined as L=L1 + λ1 Ladv + λ2 Lpercep, where we empirically chose λ1 = 0.1 and λ2=0.5.
1. Christen, T., Gong, E., Guo, J., Moseley M., &Zaharchuk G. (2018). Predicting Contrast Agent Enhancement with Deep Convolution Networks. ISMRM 2018 Proceedings.
2. Gong, E., Pauly, J. M., Wintermark, M., & Zaharchuk, G. (2018). Deep learning enables reduced gadolinium dose for contrast‐enhanced brain MRI. Journal of Magnetic Resonance Imaging.
3. Oktay, O., Schlemper, J., Folgoc, L. L., Lee, M., Heinrich, M., Misawa, K., ... & Glocker, B. (2018). Attention U-Net: Learning Where to Look for the Pancreas. arXiv preprint arXiv:1804.03999.
4. Isola, P., Zhu, J. Y., Zhou, T., & Efros, A. A. (2017). Image-to-image translation with conditional adversarial networks. arXiv preprint arXiv:1611.07004.
5. Johnson, J., Alahi, A., & Fei-Fei, L. (2016, October). Perceptual losses for real-time style transfer and super-resolution. In European Conference on Computer Vision (pp. 694-711). Springer, Cham.