2598

Deep learning based multi-organ detection: a step closer to liver T1 map quality assurance      
Charles E Hill1,2, Ferenc Emil Mozes2, Liam AJ Young2, Luca Biasiolli2, Matthew D Robson3, and Vicente Grau1
1Engineering Science, University of Oxford, Oxford, United Kingdom, 2Oxford Centre for Magnetic Resonance Research (OCMR), University of Oxford, Oxford, United Kingdom, 3Perspectum Diagnostics LTD., Oxford, United Kingdom

Synopsis

Quantitative Liver T1 mapping techniques are becoming more prevalent in the clinic, for the diagnosis and prognosis of various liver diseases. However, time pressure and technique complexity can often lead to incorrect acquisition due to sub-optimal slice location. We implement a faster Region Proposal CNN for the detection of multiple organs in the body, and report back the findings promptly via a local user interface, to inform the radiographers if they need to re-acquire. We show that this network has a high localisation accuracy and high speed, such that it will be applicable in the clinical environment.

Introduction

Quantitative T1 mapping techniques are increasing in popularity in the clinic across the world, due to their ability to correlate well with liver fibrosis and clinical outcomes.1,2 However, the diagnostic performance of T1 maps depends on how carefully measurements were performed by scanner operators. In the case of liver T1 mapping, a large cross-sectional area of the liver should be imaged to reduce error in the measurement. Unfortunately, due to strict time pressures in clinical situations, minimal time is spent choosing the optimal slice location.3 This can lead to an insufficient area of liver being imaged and analysed, which reduces their utility in aiding clinical decisions, therefore requiring re-acquisition. Current checks are often performed too late to allow for immediate correction.

In this work, we implement a faster-region convolutional neural network (fRCNN) for organ detection in quantitative T1 mapping, to identify and localise multiple organs, determining in this way if the optimal slice has been selected. The model was tested in-situ on the scanner, to determine if it performs on a timescale that is useful for radiographers to perform an immediate re-acquisition.

Methods

An fRCNN with a MobileNetV2 backbone was used (Fig. 1) for bounding box detection of the body, liver, lungs and heart.4,5 The input was a 288 x 384 x 3 channel image of the liver, where the 3 channels correspond to the 1st, 4th and 7th inversion time (TI) of the shMOLLI T1 mapping technique, such that we give the network information about the T1 recovery curve, but reduce computational intensity that would come with the passing all of the acquisitions. The network, pre-trained on the ImageNet Classification Challenge, was trained on 191 T1 acquisitions of the Liver from the UKBiobank dataset, which had been manually annotated using purpose built software, for 2000 epochs.6 At each epoch the network was validated on 34 acquisitions. During training we minimised the multi-task loss as described by Girshick et al. and monitor the validation loss for overfitting.7 Training was performed on a NVIDIA GeForce GTX 1060 6GB GPU.

After training, we output bounding boxes for 25,000 acquisitions from the UKBiobank, and randomly selected 100 of these for result-blind annotation by two MR physicists (MRP1 and MRP2), to form a test set. An intersect over union (IoU) cut-off of 0.5 was deemed a successful localisation. A receiver operating characteristic (ROC) curve and precision recall (PR) curve were used to evaluate the network's performance. As the network predicted multiple bounding boxes classes, we flattened all the predictions and quote the micro-average ROC and PR of the network, which is a measure of the whole network performance that accounts for class imbalance.

In addition, we validated the usefulness of the network by running the model in real-time on the scanner for 5 healthy volunteers. For each volunteer, the shMOLLI T1 mapping sequence was run 10 times and the time taken from data acquisition, to the feedback interface (Fig. 2) being updated, was recorded. The time taken for the model to run independently of the interface, was also assessed. The model was run on an Intel Xeon E-2124G CPU.

Results

Output bounding boxes for a subset of subjects are displayed in Fig. 3. The mean IoU between the 1st annotator and the network for each class are displayed in Table 1, alongside the mean IoU between the two annotators. All classes predicted by the network had a mean IoU greater than 0.5, with the body and liver greatly exceeding that. The network had a high performance with an area under the ROC curve (AUC) of 0.96 and a mean Average Precision (mAP) of 0.95 (Fig. 4)

Time studies on healthy volunteers showed that the network took on average 3.24 ± 0.18 s to process one set of acquisitions. The whole framework, from the output of DICOM files from the scanner, to interface update, took 3.55 ± 0.20 s to complete.

Discussion

The proposed fRCNN identifies the body, liver, heart and lungs with a high classification accuracy as shown by the ROC and PR curves. Additionally, it achieved a good localisation accuracy on the body and liver as shown by the high IoU with the MR physicists. This can be used to determine if a sufficient area of liver is in the slice. There was good agreement between the two annotators, with the exception of the lungs. This is potentially due to one annotator labelling sub sections of lung tissue with respect to the side they belonged to, and the other labelling each piece of lung tissue independently. The whole framework runs on what we deem to be on a useful timescale, as it will not unduly extend the clinical workflow.

Conclusions

In this study, we implemented an object detection network to detect and localise the body, and various organs within it in an axial slice with good success. We then showed that this network outputs bounding boxes that can be displayed intuitively, on a timescale that will not extend scanning time unnecessarily. Future work will look at further detection of smaller artefacts within the liver and their effect on quantitative parameters.

Acknowledgements

This work was supported by funding from the Engineering and Physical Sciences Research Council (EPSRC) and Medical Research Council (MRC) [grant number EP/L016052/1].

References

1. Pavlides, M. et al. Multiparametric magnetic resonance imaging predicts clinical outcomes in patients with chronic liver disease. J. Hepatol. 64, 308–315 (2016).

2. Banerjee, R. et al. Multiparametric magnetic resonance for the non-invasive diagnosis of liver disease. J. Hepatol. 60, 69–77 (2014).

3. Beardmore, C., Woznitza, N. & Goodman, S. The Radiography Workforce Current Challenges and Changing Needs. Coll. Radiogr. 1–4 (2016).

4. Ren, S., He, K., Girshick, R. & Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 39, 1137–1149 (2017).

5. Sandler, M., Zhu, M., Zhmoginov, A. & Mar, C. V. MobileNetV2: Inverted Residuals and Linear Bottlenecks.

6. Russakovsky, O. et al. ImageNet Large Scale Visual Recognition Challenge.

7. Girshick, R. Fast R-CNN. Proc. IEEE Int. Conf. Comput. Vis. 2015 Inter, 1440–1448 (2015).

Figures

Fig 1. The implemented network architecture. Images of size 288 x 384 x 3 are input into the MobilenetV2 backbone, which outputs 1280 feature maps. These feature maps are used to generate region proposals which are subsequently combined with the feature maps at different aspect ratios. These proposals are then classified into the output classes. The layers in the backbone are written as m(h x w x c), where m is the number of blocks, h and w are the dimensions of the layer, and c is the number of channels output.

Fig 2. Screenshot of the Feedback interface for the network. It displays the most recent acquisitions with bounding boxes annotated. It then provides a judgement on whether a re-scan is required based on the organs present. This judgement is advisory, and the radiographer then chooses whether or not to re-acquire.

Fig 3. A Subset of cases output from the network. A and B, depict cases with all classes present, with B under performing on lung detection. C and D, depict the majority of cases where only the Liver (and Body) are imaged. E and F, depict the minority where no liver has been imaged at all. [Liver – Green, Body – Red, Heart – Yellow, Lungs – Magenta]

Table 1. The mean IoU for each class, between the two MR Physicists (MRP1 and MRP2), and between MRP1 and the Network predictions. There is a high bounding box localisation correlation between MRP1 and MRP2, and between MRP1 and the model. A lower performance is seen in the case of lungs and the heart by the network, likely due to a lower proportion of training data with those cases (< 5%).

Fig 4. A Receiver Operating Characteristic (ROC) curve and Precision-Recall (PR) curve depicting the networks performance. The pink curve is a micro-average curve which depicts overall performance of the network, which takes into account any class imbalance. The faded lines underneath depict the performance of the individual classes, in a one versus all analysis. The Body is the best individual performer, with Lungs being the worst

Proc. Intl. Soc. Mag. Reson. Med. 28 (2020)
2598