3812

Deep Learning based spine labeling with three-plane 2D localizers without vertebrae segmentation
Dattesh Dayanand Shanbhag1, Arathi Sreekumari1, Soumya Ghose2, Chitresh Bhushan2, and Uday Patil3
1GE Healthcare, Bangalore, India, 2GE Global Research, Niskayuna, NY, United States, 3General Electric Company, Bangalore, India

Synopsis

In this work , we describe a deep learning-based methodology to generate vertebrae labels directly from the standard 2D tri-planar localizer images without the need any additional scanning or explicitly segmenting the vertebrae. This is accomplished by using deep-learning setup a to identify vertebrae labels directly on the localizer images. The method is demonstrated on lumbar spine localizer data to identify Thoracic-12 (T12), Lumbar-4 (L4) , and Sacral-1 (S1) vertebrae locations. In a test cohort of 50 lumbar MR spine exams, we report labeling accuracy of 92%, 98% and 96% for T12, L4 and S1 vertebrae respectively on localizer images.

Introduction

Spine labeling is an important task in planning and reporting of MR spine exams in routine clinical practice. During planning stage, MR technologist needs to label the vertebrae on localizer images, so that a scan order can be completed. Typically, MR spine labeling on localizer images is done manually.
Automated methods exist, but they rely either on the higher resolution 2D images or specialized 3D scouts with isotropic resolution [1] for labeling. This disrupts the clinical workflow since additional scans must be done for this purpose. The standard image processing approach is to generate the vertebrae segmentation masks and then do labeling with initial seed [2] or additionally predict the vertebrae labels using machine learning [3].
In this work , we describe a deep learning (DL)-based methodology to generate vertebrae labels directly from the standard 2D tri-planar localizer images without the need for any additional scanning or explicitly segmenting the vertebrae. This is accomplished by using deep-learning based setup a to identify vertebrae points directly on the localizer images. The method is demonstrated on lumbar spine localizer images to identify Thoracic-12 (T12), Lumbar-4 (L4) , and Sacral-1 (S1) vertebrae labels.

Methods

Subjects: Spine MRI data for the study came from a single site. A total of 122 Spine exams from clinical subjects were included in the study. All the studies were approved by appropriate IRB.
MRI Scanner and Acquisition: Regular 2D, three-plane single shot fast spin echo (SSFSE) localizer data was acquired on 1.5T MRI scanner (GE OptimaMR350), with Spine Array coil. While the data was acquired for all the three spine stations (cervical, thoracic and lumbar), we included only the data from lumbar station in this study. Acquisition parameters : TE/TR = 37.7/876 ms, Acquisition matrix = 288x192, recon matrix = 512x512, slice-thickness = 5 mm, slices variable across subjects.
Ground-truth (GT) generation: A trained radiologist marked the locations of T12, L4 , and S1 vertebrae locations on the sagittal localizer images using the ITK-Snap tool [4]. The size of the brush was fixed to square 18 units for each label [Figure 1]. The radiologist had access to higher resolution whole spine sagittal T2 image for reference purposes. L4 and S1 vertebrae were marked in every case, while T12 vertebrae was not necessarily available in each case.
DL Architecture: Deep learning CNN based 2D UNet architecture [5] with size-weighted dice loss was adapted for multi-label segmentation of vertebrae location marking (See Figure 2). Shape consistency was implemented to ensure square shape for segmented labels. All experiments done using Keras package (v2.2.4) with TensorFlow backend (v1.15.0).
Deep-Learning Data: 72 cases were chosen for DL model training and 50 cases for testing the algorithm. The training datasets were further augmented using rotations, translations and image resampling, generating 3155 slices for training, with 10% of training cases used for validation and model selection. Finally, the image and associated label data were resampled to a grid size of 128x128 and input image z-score normalized for training purposes. During model inferencing, the pre-processing steps were replicated, model inferred and predicted multi-label data resampled to native image for computing assessment metrics.
Accuracy Assessment: Localization error was computed as Euclidean distance error between the centroids of the ground-truth and DL -predicted vertebrae labels. Since vertebrae is ~ 16-18 mm in length and breadth, centroid distance error < 8 mm was considered acceptable for labeling purposes.

Results and Discussion

Figure 3 shows the scatterplot for localization errors in T12, L4 and S1 locations in 50 test cases, with the acceptability cutoff line indicated at 8mm. The network predicted T12 with mean localization error = 2.4 mm ± 2.6 mm, L4 location with a mean error = 3.2mm ±4.5mm, S1 with mean error = 2.7 mm ± 4.1 mm and are within the acceptable limit of 8 mm. Figure 4 shows the sample results in cases with good performance and those with outliers. For T12 vertebrae location, the algorithm matched with GT for absence of vertebrae location in 6/9 cases. In two cases, DL predicted a label when GT indicated it missing (false-positive) while vice-versa (false-negative) in one case. The outliers in the data are not consistent across labels in a particular dataset. For e.g. for case with L4 error = 30 mm, the error for S1 was 1.64 mm. Similarly, for case with S1 error = 23 mm, the error for L4 was only 1.3 mm.
Using a cutoff of 8 mm and accounting for false-positive and false-negative predictions, overall success rate for label accuracy was: T12 = 92%, L4 = 98% and S1 = 96%.
The results indicate that by using a denser set of labels, the error will potentially be reduced further since closer vertebrae labels will act as anchor points for neighboring labels. The results also suggest that label error in one vertebra can be compensated with other vertebrae labels markings as well.

Conclusion

We have demonstrated DL based methodology for labeling vertebrae on regular tri-planar localizers. The method has acceptable performance on clinical data, reduced manual labeling effort and thereby potentially adapted for spine exam planning workflow by technologist. Next, we plan to extend the work to whole spine localizers using dense labeling.

Acknowledgements

No acknowledgement found.

References

1. “Robust MR Spine Detection Using Hierarchical Learning and Local Articulated Model”, Yiqiang Zhan et.al., MICCAI 2012, Part I, LNCS 7510, pp. 141–148, 2012.

2. “A Deep Learning Based Solution for Vertebrae Segmentation of Whole Spine MR Images: A step closer to Automated Whole Spine Labeling”, Kavitha Manickam et.al., Proceedings of the Joint Annual Meeting of ISMRM-ESMRMB, Paris, 2018, p. 3143

3. “Automatic Localization and Labelling of Spine Vertebrae in MR Images Using Deep Learning.” naei, Pardiss. 2020. Electronic Theses and Dissertations (ETDs) 2008+. T, University of British Columbia. doi:http://dx.doi.org/10.14288/1.0390680.

4. Paul A. Yushkevich, Joseph Piven, Heather Cody Hazlett, Rachel Gimpel Smith, Sean Ho, James C. Gee, and Guido Gerig. User-guided 3D active contour segmentation of anatomical structures: Significantly improved efficiency and reliability. Neuroimage 2006 Jul 1;31(3):1116-28.

5. “U-Net: Convolutional Networks for Biomedical Image Segmentation”, Olaf Ronneberger, Philipp Fischer, Thomas Brox, Medical Image Computing and Computer-Assisted Intervention (MICCAI), Springer, LNCS, Vol.9351: 234--241, 2015

Figures

Figure 1. Ground-truth marking in representative datasets of lumbar station sagittal localizer images used for DL training. There is a wide variation in spine shape and curvature across images. In last image, the T12 is missing and hence not labeled.

Figure 2. Deep learning architecture for vertebrae labeling

Figure 3. Centroid error for 50 test cases for T12, L4 and S1 labeling. Cutoff for acceptable performance ( 8 mm ) is shown in dotted lines. Most of the errors are within the threshold limits. Overall, accuracy for T12, L4 and S1 is 92%, 98% and 96% respectively. Arrows indicate cases C and D shown in Figure 4 for L4 and S1 error respectively.

Figure 4. Sample examples of Spine labeling on 2D localizers for four different test subjects. DL predictions are shown as Red square for T12, green square for L4 and blue square for S1 respectively. A and B : Examples of cases with error much less than acceptable criteria ( < 8 mm error). C: Example with large L4 error and D: Example with large S1 error. Notice that error in one vertebra labelling in C and D doesn’t not affect the accuracy of other labels.

Proc. Intl. Soc. Mag. Reson. Med. 29 (2021)
3812