0829

CycleSeg-v2: Improving unpaired MR-to-CT Synthesis and Segmentation with pseudo label and LPIPS loss
Huan Minh Luu1, Gyu-sang Yoo2, Won Park2, and Sung-Hong Park1
1Bio and Brain Engineering, Korea Advanced Institute of Science and Technology, Daejeon, Korea, Republic of, 2Department of Radiation Oncology, Samsung Medical Center, Seoul, Korea, Republic of

Synopsis

Radiotherapy treatment typically requires both CT and MRI as well as labor intensive contouring for effective planning and treatment. Deep learning can enable an MR-only workflow by generating synthetic CT (sCT) and performing automatic segmentation on the MR data. However, MR and CT data are usually unpaired and limited contours are available for MR data. In this study, we proposed CycleSeg-v2 that extends the previously proposed CycleSeg to work with unpaired data. To ensure robust training, we employed LPIPS loss in addition to pseudo label. Experiments with data from prostate cancer patients showed that CycleSeg-v2 improved upon previous approaches.

Introduction

Radiotherapy treatment planning typically requires both CT and MRI for identifying the tumors and the surrounding organs as well as for radiation dose calculation. Manual, time-consuming contouring of relevant organs is also necessary for target delineation and treatment monitoring. Deep learning can generate both realistic synthetic CT (sCT)1,2 and segmentation3,4 from MR images to enable MR-only radiotherapy treatment planning, reducing the burden for both patients and clinicians. However, contouring is typically performed on the CT images, so often no contour is available for MR images. Since MR and CT images are not acquired at the same time, previous approaches have used CycleGAN for unpaired training to generate sCT from MR images. This can introduce geometric distortion between the MRI and sCT images as the cycle consistent training alone does not penalize this error15. In this study, we proposed CycleSeg-v2, a CycleGAN-based network that performs both sCT synthesis and segmentation based on unpaired CT and MRI. To address the aforementioned obstacles, CycleSeg-v2 incorporates the Learned Perceptual Image Patch Similarity (LPIPS) loss5 with the pseudo-label approach15 to train with unpaired data.

Methods

Data:
We utilized 115 sets of simulation CT scans of patients with prostate cancer. The simulation CT scans were performed with GE Healthcare Discovery CT590 RT. The field of view (FOV) was 50cm, the slice-thickness between the consecutive axial CT images was 2.5 mm, and the matrix size and resolution were 512×512 and 0.98×0.98 mm2. Contours of organs-of-interest were manually annotated on the CT data. MRI scans were conducted immediately after the simulation CT scans with a Philips 3.0 T Ingenia MR. T2 weighted MR images were obtained with FOV = 400×400×200 mm3, matrix size = 960×960×80, TR/TE = 6152ms/100ms. The data were manually inspected to remove those with unusual artifacts, resulting in 93 pairs of MRI/CT volumes. To process the data for training the networks, both MRI and CT volumes were resampled to a common 1.3×1.3×2.5 mm3 spacing and registered the MRI volume to CT volume with affine transformation in simpleITK6-8. The central 352×352 pixels in the axial plane were cropped to remove the background. To perform intensity normalization, the Hounsfield units for CT were linearly mapped from [-1024, 3071] to [-1, 1] range. The intensities of the MRI data were normalized by subtracting the mean, dividing by 2.5 times the standard deviation of the voxels inside the body, and clipping the normalized values to [-1,1] range. During training, MR and CT slices were randomly sampled from all training subjects.
Networks:
CycleSeg-v2 consists of a CycleGAN9 and 2 UNets10, whose structures are shown in Figures 1 and 2. Similar to Zhang et al11 and Luu et al15, CycleSeg-v2 enforces shape consistency through the segmentation network to improve the synthesis. CycleSeg-v2 adopted the pseudo-label approach proposed in CycleSeg15 to train the UNet-CT with sCT by using the masks generated by UNet-MRI from input MR images. To reduce the geometric distortion from the unpaired training scheme, we utilized the LPIPS loss, which minimizes the perceptual similarity distance using a pretrained network (VGG16). The following loss is minimized during the training process:

$$\mathcal{L}(x_{MRI},x_{CT})=\mathcal{L}_{LSGAN}+10\mathcal{L}_{cycle}+\mathcal{L}_{seg}+\mathcal{L}_{LPIPS} (1)$$
CycleSeg-v2 was compared with the baseline CycleGAN as well as SynSegNet12. To investigate the effect of the LPIPS loss, CycleSeg-v2 was compared with CycleSeg, which does not have the LPIPS loss, as well as a variant of CycleSeg with the gradient correlation loss13. Mean absolute error (MAE) of HU and Frechet inception distance (FID)14 were used as quantitative metrics for sCT synthesis and 3D dice coefficient was used to evaluate segmentation performance.

Result and Discussion

Table 1a and Figure 3a showed the effect of the LPIPS loss to enforce consistency between the input and the generated images. The addition of the LPIPS loss in CycleSeg-v2 leads to lower MAE and FID (105.22 and 12.87) compared to the cases with no LPIPS loss (182.80 and 15.28) or with the gradient correlation loss (135.92 and 14.60). Without any consistency loss, CycleSeg tends to produce sCT images that are slightly misaligned with respect to the input MR images. With the addition of either the gradient correlation or the LPIPS loss, this problem was largely negated but CycleSeg-v2 was better at generating correct CT intensity (white arrows). Table 1b compares the performance of CycleSeg-v2 with other approaches for generating sCT. It had the lowest FID value of 12.87 and the second-lowest MAE value of 105.22. Figure 3b showed a representative real MR and CT slice with generated sCT and segmentation from the 3 methods. The vanilla CycleGAN method showed the misalignment problem and failed to segment the bladder. Both SynSegNet and CycleSeg-v2 showed better synthesis and segmentation performance, with slightly more accurate right femoral and rectum segmentation for CycleSeg-v2. Figure 4 showed the quantitative Dice scores for the 3 networks in comparison to the best-case scenario or reference of Unet trained and tested on real CT. CycleSeg-v2 showed better performance compared to the other 2 networks.

Conclusion

CycleSeg-v2 incorporates LPIPS loss and pseudo label to enable simultaneous sCT synthesis and segmentation from MR images with unpaired training and no MR segmentation label. The proposed method has the potential to streamline radiotherapy treatment planning with only MR acquisition.

Acknowledgements

No acknowledgement found.

References

1. Liu Y, Lei Y, Wang Y, Shfai-Erfani G et al. “Evaluation of a deep learning-based pelvic synthetic CT generation technique for MRI-based prostate proton treatment planning”. Phys. Med. Biol. 64 205022, 2019

2. Liu Y, Lei Y, Wang T, Kayode O et al. “MRI-based treatment planning for liver stereotactic body radiotherapy: validation of a deep learning-based synthetic CT generation method”. Br J Radiol. 2019 Aug;92(1100):20190067

3. S. Elguindi et al., "Deep learning-based auto-segmentation of targets and organs-at-risk for magnetic resonance imaging only planning of prostate radiotherapy," Phys Imaging Radiat Oncol, vol. 12, pp. 80-86, Oct 2019.

4. M. H. F. Savenije et al., "Clinical implementation of MRI-based organs-at-risk auto-segmentation with convolutional networks for prostate radiotherapy," Radiat Oncol, vol. 15, no. 1, p. 104, May 11 2020.

5. Zhang R, Isola P, Efros A, Shechtman E, Wang O, “The Unreasonable Effectiveness of Deep Features as a Perceptual Metric,” CVPR 2018.

6. R. Beare, B. C. Lowekamp, Z. Yaniv, "Image Segmentation, Registration and Characterization in R with SimpleITK", J Stat Softw, 86(8), 2018.

7. Z. Yaniv, B. C. Lowekamp, H. J. Johnson, R. Beare, "SimpleITK Image-Analysis Notebooks: a Collaborative Environment for Education and Reproducible Research", J Digit Imaging., 31(3): 290-303, 2018.

8. B. C. Lowekamp, D. T. Chen, L. Ibáñez, D. Blezek, "The Design of SimpleITK", Front. Neuroinform., 7:45, 2013.

9. Zhu J-Y, Park TS, Isola P, Efros A. "Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks", in IEEE International Conference on Computer Vision (ICCV), 2017.

10. Ronneberger O, Fischer P, Brox T. “U-Net: Convolutional Networks for Biomedical Image Segmentation”, Cham, Springer International Publishing, 2015.

11. Zhang Z, Yang L, Zheng Y. "Translating and Segmenting Multimodal Medical Volumes with Cycle- and Shape-Consistency Generative Adversarial Network." In Conference on Computer Vision and Pattern Recognition (CVPR), 2018

12. Huo, Y., Xu, Z., Moon, H., Bao, S., Assad, A., Moyo, T.K., Savona, M.R., Abramson, R.G., Landman, B.A.: SynSeg-Net: Synthetic Segmentation Without Target Modality Ground Truth. IEEE Transactions on Medical Imaging 38, 1016-1025 (2019)

13. Hiasa, Y., Otake, Y., Takao, M., Matsuoka, T., Takashima, K., Carass, A., Prince, J.L., Sugano, N., Sato, Y.: Cross-Modality Image Synthesis from Unpaired Data Using CycleGAN. pp. 31-41. Springer International Publishing, (2018)

14. Heusel M, Ramsauer H, Unterthiner T, Nessler B et al. “GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium”, in Advances in Neural Information Processing Systems (NIPS), 2017

15. Luu H.M et al., “CycleSeg: MR-to-CT synthesis and segmentation network for prostate radiotherapy treatment planning”. 30th Annual Meeting, ISMRM 2021. Program number 0813.

Figures

Figure 1: Architecture of CycleSeg-v2, which is a combination of CycleGAN and 2 Unets. The notations in this figure correspond to those in Equation (1) in the text.

Figure 2: Architectures of the generator and discriminator used in this study. The number at the top and bottom of the feature maps indicates the number of channels and spatial dimensions (width, height), respectively.

Table 1: sCT synthesis performance comparison. a) Comparison of CycleSeg-v2 with CycleSeg and CycleSeg with gradient correlation loss b) Comparison of CycleSeg-v2, CycleGAN, and SynSegNet. Lower values of FID and MAE are better. MAE is calculated between sCT and matching reference CT volume for each test subject. FID is calculated by comparing the distribution of Inception-V3’s output before the last pooling layer of all sCT and matching real CT slices.

Figure 3: Qualitative results for sCT synthesis and segmentation (displayed as colored contours). a) The effect of the LPIPS loss. Without LPIPS loss, CycleSeg (top) produces sCT images that are translated from the input images. Adding gradient correlation loss (middle) and LPIPS loss (bottom) fixes the problem but CycleSeg-v2 has more accurate HU intensity values. b) comparison of CycleSeg-v2 with CycleGAN and SynSegNet. Both CycleSeg-v2 and SynSegNet produce realistic sCT and segmentation in comparison with CycleGAN. CycleSeg-v2 was slightly more accurate for some organs.

Figure 4: Quantitative comparison of the dice score between different models. Real CT denotes results from a Unet that was trained and tested on real CT data. CycleGAN denotes results of the aforementioned Unet tested on CycleGAN sCT. Synsegnet denotes results of Unet trained on synthetic MRI data tested on real MRI data. Cycleseg-v2 denotes result of Unet trained on real and synthetic CT data tested on sCT data. CycleSeg-v2 showed better segmentation for all organs compared to CycleGAN and SynSegNet.

Proc. Intl. Soc. Mag. Reson. Med. 30 (2022)
0829
DOI: https://doi.org/10.58530/2022/0829