1869

Fast prediction of whole-brain cerebral microbleed masks from 7T SWI imaging with a deep residual 3D UNet
James R Golden1, Yicheng Chen1, Melanie A Morrison1, Kate Nelson1, and Janine M Lupo1
1Radiology, UCSF, San Francisco, CA, United States

Synopsis

Cerebral microbleeds (CMBs) are known risk factors of stroke and hemorrhage that can be a marker of cognitive impairment. Although CMBs are easily visualized with susceptibility weighted imaging, they are burdensome to localize and quantify manually. Image processing algorithms based on the radial symmetry transform have previously been used to identify candidate CMBs, and convolutional neural networks have been effective at distinguishing real CMBs from mimics with high sensitivity and specificity. A deep neural network was trained to carry out this entire pipeline and to predict CMB voxel masks using a dataset of radiation therapy-induded CMBs from patients with gliomas.

Introduction

Cerebral microbleeds (CMBs) are known risk factors of stroke and hemorrhage that can be a marker of cognitive impairment. Although CMBs are easily visualized with susceptibility weighted imaging (SWI), they are burdensome to localize and quantify manually. Image processing algorithms based on the radial symmetry transform have previously been used to identify candidate CMBs1,2, and convolutional neural networks (CNNs) have been effective at distinguishing real CMBs from their false positive mimics with high sensitivity and specificity3,4. The goal of this study was to train a deep neural network to carry out this entire processing pipeline and to predict CMB voxel masks using a dataset of radiation therapy-induced CMBs from patients with gliomas.

Methods

Dataset: 7T SWI scans of brain tumor patients with radiation induced CMBs were acquired on a GE 7T scanner equipped with a 32-channel receive coil using either a standard SWI sequence (resolution = 0.5x0.5x2mm, FOV. = 24x24cm, TE = 16ms, TR = 50ms, 49 subjects) or a 4-echo 3D TOF-SWI sequence5 (resolution=0.5x0.5x1mm, TE = 2.7, 10.4, 13.2, 20.9 ms, TR = 40ms, 31 subjects). Figure 1b shows the breakdown of the number of patients, exams, and CMBs in each of the training, validation, and test datasets.

Pre-processing: To acquire a labeled dataset to pre-train the network, the SWI scans were initially processed with a fast radial symmetry transform algorithm1 to find candidate CMBs, which were then inspected using a user-friendly GUI2 before segmenting the resulting candidate CMBs and false-positive mimics. The center coordinates of these labeled masks were then used to extract 16 x 16 x 8 3D patches to use as inputs to the ResNet CNN as shown in Figure 1a, with the target output of 1 corresponding to a true CMB patch. For the mask segmentation training, the same CMB and mimic input patches were used, along with their corresponding segmented labeled masks as target outputs.

Deep network model strategy: A ResNet architecture with 3D convolutional kernels3 followed by fully connected blocks was first trained to classify voxel patches as containing either a CMB or a mimic feature. The residual blocks from this network were then used as the pre-trained front end for a UNet architecture that was subsequently trained to generate segmented voxel masks for CMB patches as shown in Figure 2. The network was trained using a Nvidia Titan Xp GPU with 12GB memory and the Adam optimizer that began with a learning rate of 0.001 and experienced a cosine decay over each epoch for ~100,000 iterations. A batch size of 32 was utilized and a weighted binary cross entropy was employed as the loss function, where the weight was equal to the number of voxels that were part of a CMB divided by the total number of voxels in the training set. In order to avoid overfitting, data augmentation including random rotations and flips was carried out.

Results and Discussion

The classification network used in pre-training achieved an ROC area under curve (AUC) of 97.7% and an average precision of 93.5%. Figures 3 shows the Receiver Operating Characteristic (ROC) and Precision-Recall curves for the results of the mask segmentations. The mask prediction network achieved a dice score of 0.755 on patched-based predictions, but only 0.644 when given the whole brain at once in testing. The corresponding confusion matrices are shown in Figure 4. A comparison of the network predictions for two representative CMB and mimic masks for two extreme cases with very large CMB/mimic features is illustrated in Figure 5.

After the UNet has been trained to predict CMB voxel masks, it can still be used to classify patches by checking to see if any voxel in a patch has a predicted likelihood of being a CMB above a threshold. Using this method, the ROC AUC is 95.7% and the average precision is 84.8%, with only 12.6 false-positive CMB mimics per patient. Although the classification performance was reduced compared to the original classification architecture, the mask prediction task is different because it must learn 3D position information about CMBs and it requires a much larger network to reach the same performance standard as the classifier. Including a second input channel corresponding to the phase image4 during training did not improve classification or mask prediction performance.

The whole brain prediction utilizes test-time augmentation where the input data is flipped and rotated in a number of ways such that 32 different predictions for the same brain are averaged together. This augmentation provided a benefit in mask prediction precision but extends the total computation time on a Titan XP GPU to 80 seconds as opposed to 30 seconds without augmentation.

Conclusion

In spite of the apparently obvious appearance of CMB features on SWI images, segmenting accurate CMB masks that exclude mimic features is difficult. We trained a UNet architecture to predict the positions of CMB voxels on a large dataset of SWI patches, and demonstrated high sensitivity and specificity on test sets of patch-based inputs. The high performance and short computation time of the network for automating CMB quantification point toward a possible clinical application.

Acknowledgements

Funding Information: this work was supported by the National Institute for Child Health and Human Development of the National Institutes of Health grant R01HD079568 and GE Healthcare.

References

1. Bian, W., et al. "Computer-aided detection of radiation-induced cerebral microbleeds on susceptibility-weighted MR images." NeuroImage: clinical 2 (2013): 282-290.

2. Morrison, M., et al. "A user-guided tool for semi-automated cerebral microbleed detection and volume segmentation: Evaluating vascular injury and data labelling for machine learning." NeuroImage: Clinical 20 (2018): 498-505.

3. Chen, Y., et al. “Toward Automatic Detection of Radiation-Induced Cerebral Microbleeds Using a 3D Deep Residual Network.” Journal of Digital Imaging. (2018)

4. Liu, Saifeng, et al. "Cerebral microbleed detection using Susceptibility Weighted Imaging and deep learning." NeuroImage 198 (2019): 271-282.

5. Bian, W., et al. “Simultaneous imaging of radiationā€induced cerebral microbleeds, arteries and veins, using a multiple gradient echo sequence at 7 Tesla.” Journal of Magnetic Resonance Imaging, 42.2 (2015): 269-279.

Figures

Figure 1: a) The SWI scans are initially evaluated by the radial symmetry transform algorithm. Candidate patches are labeled by humans using a GUI, allowing patches to be extracted and used to training a deep network for either classification or mask prediction. b) The dataset consists of scans from 73 patients with glioma broken into training, validation and test sets.

Figure 2: ResNet and UNet designs. The residual blocks from the ResNet are pretrained on the classification task and used as the front end of the UNet. The ResNet had about 12M parameters and the UNet had 24M parameters. Although a much smaller network is sufficient for the classification task, a ResNet of this size was needed to maximize performance in the mask prediction task.

Figure 3: The output represents the likelihood that the input voxel is part of a CMB, so the predictions can be assessed as classifications at the voxel level. The dice coefficient is used for accuracy of image segmentation, and is computed as 2*TP/(2*TP+FP+FN). The ROC area under curve (AUC) is not particularly informative here because of the imbalance of CMB voxels to non-CMB voxels (1 CMB voxel for every 600 non-CMB voxels), which creates a large number of true negative (TN) classifications.

Figure 4: Example confusion matrices for the a) patch and b) whole-brain input. The network performs worse for the whole-brain input, most likely due to the diversity of non-CMB features that were not found in the training set. Note that the precision-recall curves in Figure 3 better capture the performance of the network over all possible confusion matrices.

Figure 5: A comparison of the network predictions for CMB voxels and mimic voxels for two extreme cases with very large CMB/mimic features. Note that the predictions differ at many edge voxels from the hand labeling.

Proc. Intl. Soc. Mag. Reson. Med. 28 (2020)
1869