1291

Connecting Histology and MRI using Deep Learning
Zifei Liang1, Choong Heon Lee1, Tanzil M. Arefin1, Piotr Walczak2, Song-Hai Shi3, Florian Knoll1, Yulin Ge1, Leslie Ying4, and Jiangyang Zhang1
1Radiology, NYU Langone Health, New York, NY, United States, 2Diagnostic Radiology & Nuclear Medicine, University of Maryland School of Medicine, Baltimore, MD, United States, 3Center for Molecular Imaging & Nanotechnology, Memorial Sloan Kettering Cancer Center, New York, NY, United States, 4Electrical Engineering, University at Buffalo, Buffalo, NY, United States

Synopsis

We developed a deep learning network that can generate new tissue contrasts from MRI data to match the contrasts of several histological methods. The network was trained using the carefully curated histological data from the Allen Institute mouse brain atlas and co-registered MRI data. In our tests, the new contrasts, which resembled Nissl, neurofilament, and myelin-basic-protein stained histology, demonstrated higher sensitivity and specificity than commonly used diffusion MRI markers to characterize neuronal, axonal, and myelin structures in the mouse brain. The contrasts were further validated using two mouse models with abnormal neuronal structures and dysmyelination.

Introduction

MRI is an indispensable tool for non-invasive detection of neuropathology. The unparalleled rich tissue contrasts it provides, however, are mostly indirect measurements of tissue properties and lack specificity compared to histology. In this work, we investigated whether deep learning can generate MRI-based contrasts to match the contrasts of several histological methods. We have developed a series of deep learning networks, called MRHNet, based on the carefully curated histological data from the Allen mouse brain atlas1 and co-registered high-resolution MRI data. The MRI data consisted mainly of a large set of diffusion MRI (dMRI) data, which offer several tissue contrasts linked to tissue microstructure. The results were validated using test data from normal and genetically modified mouse brains. We further sought to understand the inner-working on the network and used the network to evaluate the contribution of individual image.

Materials & Methods

A series of deep learning networks, called MRHNet, were constructed to take multi-contrast MRI data (mostly ex vivo mouse brain dMRI data, C57BL/6, n=12) as inputs and co-registered histological images from the Allen Brain Atlas (ABA) as the target (Fig. 1A), including Nissl/Neurofilament/MBP (for neuronal/axonal/myelin structures) as the target. We used data from part of the forebrain region for training and the rests for testing. For further validation, we acquired data from the Sas4-/-p53-/- mouse brains, which had abnormal cortical cell masses2, and litter-mate controls (n=4/4) and the shiverer mouse brains, which showed dysmyelination3, and controls (n=4/4). Ex vivo dMRI data were acquired with 30-60 directions, b-value of 700-5000 s/mm2, both oscillating (OGSE, 50-150 Hz) and pulsed gradients (PGSE, Δ=15ms)4, and a resolution of 0.1x0.1x0.1 mm3 to match the ABA data. A 5X5 patch size was used to accommodate residual mismatches between histology and MRI data. Approximately 40,000 such 5X5 patches were used to train the network.

Results

We first tested whether MRHNet trained using Nissl data can generate contrasts that resembled the tissue contrasts in Nissl stained sections. In the testing regions, the network that utilized all OGSE and PGSE dMRI data as inputs generated MRHNet-Nissl maps with good agreement with the ground truth Nissl data (Fig. 1B-C), with higher sensitivity and specificity than results with partial data (Fig. 1B-D). The MRHNet-Nissl map of the Sas4-/-p53-/- mouse brains produced image contrasts that closely matched Nissl-stained histology.

We then used the t-Distributed Stochastic Neighbor Embedding (t-SNE)5 to visualize the feature space to better understand the inner-working of MRHNet. In Fig. 2A, a majority of the patches corresponded to regions with low Nissl signals, whereas patches in the lower-left corner corresponded to regions with strong Nissl signals and several patches on the upper-right corner corresponded to the brain surface. Fig. 2B shows representative signal profiles from the three categories. Patches that corresponded to strong Nissl signal regions showed decreased signals as the oscillating frequency increased, whereas the other two types of patches show no such pattern.

We also trained networks using axon and myelin stained histology (Fig. 3A-B), and both showed improved sensitivity and specificity than commonly used fractional anisotropy (FA) and radial diffusivity (DR), as shown by the ROC analysis (Fig. 3C-D) and regional difference maps (Fig. 3E). In Fig. 3F, MRHNet-MBP map based on dMRI data from the dysmyelinated shiverer mouse brain suggested reduced myelin than controls.

We further analyzed the contribution of each dMRI data to the network outcomes by replacing the individual image with an image with constant signals and measured the RMSE of the results compared to the ground truth (Fig. 4). We found that the contribution of individual dMRI data were unevenly distributed.

Discussions

Our results demonstrate that deep learning can assist the development of highly specific markers from MRI data. While significant correlations between several dMRI-based markers and histological measurements of axon and myelin have been reported, the proposed network can potentially generate the optimal markers for the given MRI data. The co-registered MRI and histological data enabled us to validate the sensitivity and specificity of these markers. The technique may assist the development of optimal multi-contrast MRI and new MRI contrasts by quantifying the contribution of individual image to detect certain histopathological features.

Our study is not without its limitations: 1) the MRI and histology data used in this study were acquired from different mouse brains of the same strain; 2) the networks were based on ex vivo mouse data and may not apply to in vivo data due to the differences between in vivo and ex vivo MRI signals; 3) the resolution of MRI remains limited compared to histological data; and most importantly, 4) data from cases with complex neuropathology, e.g., inflammation and edema, are not present in the training dataset, which may limit the applicability of the technique for such cases.

Conclusion

Our results demonstrate that deep learning based on co-registered MRI and histological data can improve the sensitivity and specificity of MRI in detecting certain histopathological features in the brain.

Acknowledgements

This study was supported by NIH R01NS102904

References

1. Allen Mouse Brain Atlas Version 2 (2011) Technical White Paper, https://mouse.brain-map.org/static/atlas.

2. Insolera, R. et al. Cortical neurogenesis in the absence of centrioles, Nat Neurosci. 2014 Nov: 17(11): 1528-1535.

3. Weil, M.T. et al. Loss of Myelin Basic Protein Function Triggers Myelin Breakdown in Models of Demyelinating Diseases, Cell Rep. 2016 Jul 12; 16(2): 314-322.

4. Aggarwal, M. et al. Probing Mouse Brain Microstructure Using Oscillating Gradient Diffusion Magnetic Resonance Imaging, Magn Reson Med. 2012 Jan; 67(1):98-109

5. Van der Maaten, L.J.P. et al. Visualizing Data using t-SNE, J Machine Learning Research. 2008 Nov; 9:2579-2605

Figures

Fig. 1: Training and validation of MRHNet-Nissl. A: Part of the co-registered dMRI and Nissl data were used for training (the purple box), and the rest for testing. B: Comparisons of the Nissl with MRHNet results with different inputs. C-D: The MRHNet-Nissl with PGSE+OGSE showed strong correlation with Nissl and outperformed others in ROC analysis. E: MRHNet-Nissl maps revealed abnormal cortical structures in a Sas4-/-p53-/- mouse brain compared to controls, in agreement with Nissl histology.

Fig. 2: t-SNE analysis of the MRHNet. A: 2D map of the feature space that separating the neuronal rich region (the lower left corner, blue) from other brain regions (the center, green) as well as patches at the brain surface (the upper right corner, yellow). B: Signal profiles from two representative patches for each of the three categories. The profiles show signals from non-diffusion-weighted (b0), PGSE and OGSE DWIs.

Fig. 3: The training and validation of MRHNet-Neurofilament (NF)/Myelin-basic-protein (MBP). A: Comparisons of NF, MRHNet-NF, and fractional anisotropy (FA). B: Comparisons of MBP, MRHNet-MBP, and radial diffusivity (DR). C-D: Both results showed strong correlation with ground truth and improved sensitivity and specificity than FA and DR. E: Differences between MRHNet results and FA/DR. F: MRHNet-MBP maps show reduced myelin in the shiverer mouse brain. The unit for DR is 10-3 mm2/s.

Fig. 4: Contributions of individual image to MRHNet-Nissl contrast. Each time, a single image (top row) was replaced by a constant image as input to MRHNet, and the result (the second row) was evaluated against the ground truth. The difference in RMSE measurement was used as the contribution of the image. It is assumed that, if an image does not contribute much to the final result, replacing it with a constant image will not cause too much deviation from the ground truth and vice versa.

Proc. Intl. Soc. Mag. Reson. Med. 28 (2020)
1291