3715

The role of training on the robustness of domain-transform manifold learning
Danyal Bhutto1,2, Bo Zhu2, Jeremiah Zhe Liu3,4, Stephen Cauley2,5, Neha Koonjoo2,5, Bruce R Rosen2,5, and Matthew S Rosen2,5,6
1Biomedical Engineering, Boston University, Boston, MA, United States, 2Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, United States, 3Google, Mountain View, CA, United States, 4Biostatistics, Harvard University, Cambridge, MA, United States, 5Harvard Medical School, Boston, MA, United States, 6Physics, Harvard University, Boston, MA, United States

Synopsis

Keywords: Machine Learning/Artificial Intelligence, Image Reconstruction

Domain-transform manifold learning is a trained reconstruction approach where care needs to be taken to appropriately represent the forward encoding model during training, including for example the numerical properties of the source sensor data, phase relationship of complex sensor data, and field-of-view to prevent artifacts arising in the reconstruction. Here, we study the role that the training corpus and the numeral properties of the training have on the performance of the reconstruction of MRI data and demonstrate reconstruction artifacts that result from inference on out-of-training-distribution data if the training data is not augmented sufficiently.

Introduction

All supervised learning approaches from classification to image reconstruction require the curation of a training dataset that leads to the robustness and generalization of a deployable model. Automated Transform by Manifold Approximation (AUTOMAP) is a deep learning framework that employs domain-transform manifold learning to reconstruct MRI images from raw sensor k-space data1. Koonjoo et. al showed that the domain specificity of the training data plays an important role. AUTOMAP trained on forward-modeled synthetic roots and synthetic vascular tree structures performed significantly better on real-world root raw MRI k-space data when compared to AUTOMAP trained on forward-encoded brain images2. To form a deployable AUTOMAP model, we need to construct a training dataset that represents the forward encoding appropriately, incorporating numerical properties of the source sensor data, phase relationship of complex sensor data, and multiple field-of-views (FOV) among other properties to prevent artifacts arising in the reconstruction. We test and demonstrate that artifacts can occur and be diminished for two properties, 1) training AUTOMAP with and without a range of image-domain FOV and 2) training AUTOMAP with and without the numerical properties of engineered adversarial noise3. Especially when reconstructing medical images, it is paramount to study the properties of the training dataset so that a model generalizes and doesn’t lead to a misdiagnosis.

Materials and Methods

We trained AUTOMAP models to study two properties of training datasets, robustness to FOV artifacts and adversarial noise. The training images were 2D T1-weighted brain MR images acquired at 3 T collected from the MGH-USC Human Connectome Project (HCP) public dataset4. For the FOV experiments, we needed to curate two datasets, one with a single FOV (random samples shown in Fig 1a) and one with multiple FOVs (random samples shown in Fig 1b). Fig 1c was used to showcase artifacts from the different models in Fig 2 and discussed in the results. To study the robustness to adversarial noise and the role of numerical properties of a dataset, the adversarial noise was generated using a gradient ascent with momentum perturbation search algorithm3. The original AUTOMAP was trained to output positive, magnitude images. The adversarial noise pushed this boundary to generate negative and complex-valued noise in the image domain. When applied in the k-space domain, the numerical properties pushed AUTOMAP beyond its reconstruction capabilities leading to artifacts. In Fig 3, we show the adversarial noise is highly structured and contains negative values. Instead of training on positive, magnitude-only images, we added randomized constant offsets (between -0.1 and +0.1) to the training images and retrained AUTOMAP. We show the results of both these models in Fig 4.

Results and Discussion

The top row in Fig 2 demonstrates what artifacts arise when a model trained on a single FOV is tested on an out-of-distribution FOV in the image domain. All conditions, without noise (Fig 2a), Gaussian noise (Fig 2b), and Poisson noise (Fig 2c), display artifacts (particularly in the dark image background region). After retraining AUTOMAP with multiple FOVs, under both noiseless and noise-applied conditions, the artifacts were significantly reduced (bottom row Fig 2 d-f). We tested and demonstrated how important the role of training data augmentations is through something as simple as incorporating multiple FOVs in the training data. In a clinical setting, these same artifacts could lead to a misdiagnosis or a delay as the clinical practitioners may not be aware where the fault occurred. We also demonstrate the role of the numerical properties of a training dataset and how artifacts introduced by adversarial noise can be avoided. Fig 4a (top row) demonstrates artifacts that arise from adversarial noise. After training a network on images that contain negative offsets due to the adversarial noise consisting of negative values, the retrained AUTOMAP model doesn’t contain the same artifacts in Fig 4b (bottom row). If adversarial noise is a concern, we demonstrate it is important to study the numerical properties of the dataset and augment it appropriately so the model can generalize in practice.

Conclusion

To train a deployable image reconstruction model, the role of training data augmentations and numerical properties of datasets can’t be underestimated. We trained two models, with a single FOVs and a range of FOVs in the image domain. The model trained on a single FOV produces artifacts on an out-of-distribution FOV image while the properly trained model doesn’t. We also studied the role of numerical properties. When adversarial noise pushes a network beyond its training dataset numerical properties, we observe artifacts. If those numerical properties are incorporated into the training dataset, the artifacts are significantly reduced.

Acknowledgements

We acknowledge support for this work from the National Science Foundation Graduate Research Fellowship under Grant No. DGE-1840990 and the NSF NRT: National Science Foundation Research Traineeship Program (NRT): Understanding the Brain (UtB): Neurophotonics DGE-1633516NSF.

References

1. Zhu, Bo, et al. "Image reconstruction by domain-transform manifold learning." Nature 555.7697 (2018): 487-492.

2. N. Koonjoo, B. Zhu, G. C. Bagnall, D. Bhutto, and M. S. Rosen, “Boosting the signal-to-noise of low-field MRI with deep learning image reconstruction,” Sci Rep-uk, vol. 11, no. 1, p. 8248, Apr. 2021, doi: 10.1038/s41598-021-87482-7.

3. Antun, Vegard, et al. "On instabilities of deep learning in image reconstruction and the potential costs of AI." Proceedings of the National Academy of Sciences (2020).

4. Fan Q, Witzel T, Nummenmaa A, Dijk KRAV, Horn JDV, Drews MK, et al. MGH–USC Human Connectome Project datasets with ultra-high b-value diffusion MRI. Neuroimage 2016; 124: 1108–1114.

Figures

Fig 1: Random image samples showcasing different FOV augmentations needed to properly train AUTOMAP. A) shows a representative random sample of images from the corpus representing a single FOV used to train AUTOMAP in Fig 2 a-c. B) shows a representative random sample of images with multiple FOVs used to train AUTOMAP in Fig 2 d-f. Image (c) is the full test image used in Fig 2 to show artifacts when FOVs are not incorporated in the training data augmentation pipeline.

Fig 2: Training data augmentations need to account for multiple fields of view (FOV). We trained two models on two datasets, one built on a single FOV (top row) and one with multiple FOV (bottom row). Under different noise conditions, without noise (a), Gaussian noise (b), and Poisson noise (c), there are artifacts when reconstructing an image with a different FOV than the training data. For AUTOMAP trained with multiple FOVs, the presence of these artifacts is significantly reduced (d-f).

Fig 3. Adversarial noise exhibits highly structured features not observed in natural statistical noise. The real (top row, a-d) and imaginary (bottom row, e-h) components of additive noise are displayed in the image domain. The engineered adversarial noise samples shown (b-d and f-h) are highly structured unlike natural noise and we observe an unnatural phase bias as manifested by the dissimilarity between the real (b-d) and imaginary (f-h) components.

Fig 4: Improved reconstruction in the presence of adversarial noise by augmenting training dataset numerical properties. A) Original AUTOMAP network reconstructing in the presence of adversarial noise samples (top row). B) Retrained AUTOMAP network (adding randomized constant offsets between -0.1 and +0.1 for previously out-of-distribution negative-valued outputs) has significantly fewer artifacts previously caused by the adversarial noise (bottom row).

Proc. Intl. Soc. Mag. Reson. Med. 31 (2023)
3715
DOI: https://doi.org/10.58530/2023/3715