Anatomy contouring is essential in quantifying the dose delivered to the prostate and surrounding anatomy after low-dose-rate prostate brachytherapy. Currently, five anatomical structures including the prostate, rectum, seminal vesicles, external urinary sphincter, and bladder, are contoured manually by a radiation oncologist. In this work, we investigated six convolutional encoder-decoder networks for automatic segmentation of the five organs. Six pretrained convolutional encoders and two loss functions were investigated. This yielded twelve different models for comparison. Results indicated that classification accuracy of convolutional encoders pretrained on the ImageNet dataset positively correlated with semantic segmentation accuracy in prostate MRI.
Sixty-seven post-implant patients were scanned with a 3D balanced steady-state free precession pulse sequence (CISS) on a 1.5T Siemens Aera scanner using a 2-channel rigid endorectal coil in combination with two 18-channel pelvic array coils [3]. The prostate, rectum, seminal vesicles (SV), external urinary sphincter (EUS), and bladder were segmented by a board certified radiation oncologist (S.F.) and a radiation oncology resident (G.L.). A total of 4999 slices were available, which were split into 3666/917/416 for training/cross validation/testing.
A deep learning application engine [4] was used to construct and train all models. Six convolutional encoder-decoder networks were constructed using six convolutional encoders pre-trained on the ImageNet dataset: VGG16 [5], VGG19 [5], DenseNet121 [6], DenseNet169 [6], DenseNet201 [6], and Xception [7]. The same decoder was used for each network (Figure 1). Resize convolutional layers in the decoder were chosen over transpose convolutions to avoid the checkerboard artifacts characteristic of transpose convolutions [8]. Each of the six networks were trained using two different loss functions: cross-entropy and Tversky loss (α=0.5, β=0.5). This yielded a total of 12 different models trained to perform anatomy segmentation.
Model training was performed on a Dell 7920 rack mounted server (operating under the Linux RedHat v7.2) with 4 K2200 GPUs connected with SLI technology. An Adam optimizer was used to train all models. The initial learning rate was set to 1*10-4 and was decayed by 20% if the validation loss didn’t improve in 3 epochs. Training was terminated when no reduction in the validation loss occurred after 10 epochs.
Overall pixel-wise classification accuracy was compared among all 12 trained models. Organ-wise volumetric segmentation accuracy was assessed by computing pixel-wise classification accuracy, dice similarity coefficient, and intersection over union.
[1] Ronneberger O, Fischer P, Brox T. “U-Net: Convolutional Networks for Biomedical Image Segmentation”, arXiv:1505.04597v1, 2015.
[2] Sanders J, Lewis G, Frank S, et al. “A fully convolutional network utilizing depth-wise separable convolutions for semantic segmentation of anatomy in MRI of the prostate after permanent implant brachytherapy”, Proc. ISMRM Workshop on Machine Learning, Pacific Grove, CA, 2018.
[3] Sanders JW, Song H, Frank SJ, et al. “Parallel imaging compressed sensing for accelerated imaging and improved signal-to-noise ratio in MRI-based postimplant dosimetry of prostate brachytherapy,” Brachytherapy;17(5):816-824, 2018.
[4] Sanders J, Fletcher J, Frank S, et al. “Deep Learning Application Engine (DLAE): end-to-end development and deployment of medical deep learning algorithms”, Proc. ISMRM Workshop on Machine Learning Part II, Washington, DC, 2018.
[5] Simonyan K, Zisserman A. “Very Deep Convolutional Networks for Large-Scale Image Recognition”, arXiv:1409.1556v6, 2015.
[6] Huang G, Liu Z, van der Maaten L, et al. “Densely Connected Convolutional Networks”, arXiv:1608.06993v5, 2018.
[7] Chollet F. “Xception: Deep Learning with Depthwise Separable Convolutions”, arXiv:1610.02357v3, 2017.
[8] Odena A, Dumoulin V, Olah C. "Deconvolution and Checkerboard Artifacts", Distill, 2016. http://doi.org/10.23915/distill.00003