2934

On Training Model Bias of Deep Learning based Super-resolution Frameworks for Magnetic Resonance Imaging
Mamata Shrestha1, Nian Wang2,3, and Ukash Nakarmi1
1Computer Science and Computer Engineering, University of Arkansas, Fayetteville, AR, United States, 2Radiology and Imaging Sciences, Indiana University School of Medicine, Indianapolis, IN, United States, 3Stark Neurosciences Research Institute, Indianapolis, IN, United States

Synopsis

Keywords: Image Reconstruction, Data Processing, Image super-resolution, model bias, deep learning

Deep Learning based image super-resolution methods are biased towards training data modeling. Generalizability of DL based super-resolution frameworks can be improved by introducing variability and diversity in the training data.

Introduction

Deep learning-based methods have been the preferred choice for solving various learning problems such as image super-resolution 1 and reconstruction in the field of medical imaging. In image super-resolution, the goal is to estimate the high-resolution image from the given low-resolution image. The image super-resolution problem can be formulated as:
$$ \textbf{y} = \phi ( \textbf{x}; \theta_{d}),$$
where $$$\phi$$$ is the degradation model with parameter $$$\theta_{d}$$$ that a desired high-resolution image $$$\textbf{x}$$$ undergoes and $$$\textbf{y}$$$ is the low-resolution image as shown in figure1.
The objective of the super-resolution problem is to find an estimate $$$ \hat{ \textbf{x}}$$$ of the high-resolution image $$$\textbf{x}$$$ by estimating the inverse of the degradation function $$$\phi^{-1}(
\cdot)$$$ such that:
$$ \hat{ \textbf{x}} = \phi^{-1} ( \textbf{y}; \theta_{r}) $$
In deep learning-based super-resolution frameworks, the inverse function is learned in a supervised fashion by minimizing some loss function $$$L(\cdot)$$$, such as mean squared error, structural similarity, etc over a given set of T training data $$$(\textbf{y}_i,\textbf{ x}_i), i = 1, 2, \cdots T$$$ such that:
$$ \hat{\textbf{x}} = \hat \phi^{-1} ( \textbf{y}; \theta_{r})$$
where $$$\hat\phi^{-1}(\cdot)$$$ is the inverse of the degradation model learned by the deep learning framework.

Problem Formulation

In MRI super-resolution, often, $$$(\textbf{y}_i,\textbf{ x}_i)$$$ do not exist because i) the degradation model $$$\phi(\cdot)$$$ in MRI is non-deterministic and ii) it is not possible to acquire exact low-resolution and high-resolution image training pairs because of inherent challenges such as shifting and motion of subject between two acquisitions and prohibitively long acquisition times. Therefore, most of the work based on deep learning mimics the degradation model $$$\phi(\cdot)$$$ using simple deterministic low-pass filtering techniques 2 such as Gaussian, Hanning, Hamming, or other degradation models such as bicubic interpolation 3, mean or median filtering. Moreover, most of the current work for image super-resolution focuses on designing new architecture, training techniques, and loss functions [1,4,5,6]. In this work, we advocate a better representation of several kinds of degradation models in data is important in order to build robust deep learning-based super-resolution frameworks and to avoid misleading performance and efficiency metrics of DL-based super-resolution frameworks.
We hypothesize that:
1. DL-based super-resolution frameworks are biased toward degradation model $$$\phi(\cdot)$$$ used to simulate low-resolution images for training DL framework. In other words $$$\hat\phi^{-1}(\cdot)$$$ is biased by $$$\phi(\cdot)$$$ used to create training image pairs.
2. Enforcing the diversity of the degradation model in training dataset improves generalization and reduces the bias of the DL super-resolution framework.

Methods

Data set and Training Sample Preparation: All experiments were carried out in compliance with the local Institutional Animal Care and Use Committee. Five Alzheimer’s disease (AD) with 5xFAD background mice brains were acquired on a 30-cm bore 9.4 T magnet (BrukerBioSpec 94/30, Billerica, MA). A 3D gradient echo (GRE) pulse sequence was performed at the spatial resolution of 25 × 25 × 25 μm3, FOV = 18.0 mm × 12.8 mm × 7.6 mm, flip angle = 45°, bandwidth (BW) = 125 kHz, and TR = 100 ms. To investigate the bias of the model, we used these images as High-Resolution images and created 6 different training sets each of size $$$T = 720 $$$ using 6 different degradation models such as Gaussian, Hamming, Hanning, Bicubic, Mean and Median Filtering to mimic low-resolution images and corresponding 6 validation sets. We train a DL framework six different times using these 6 training sets separately and saved all 6 trained DL models. We also created a training set having the same number of training samples $$$T= 720 $$$ but containing an equal proportion of examples from each of the six degradations models to investigate the performance of the DL super-resolution framework when degradation diversity is enforced in training.
Deep Learning Architecture: We used a DenseNet architecture 7 as shown in Fig (2) as our DL framework. We used a network with 5 dense blocks each block consisting of 5 convolution layers with a growth rate of 5. Convolution layers have a kernel size 3 with stride and padding equal to 1, for all experiments.

Result and Discussions

Figure 3 shows NRMSE and HFEN 8 performance metrics for 7 different experiments each representing 7 different $$$\phi(\cdot)$$$ used to train the deep learning framework. We can see when model $$$\phi(\cdot)$$$ in the training and validation set is the same, both NRMSE and HFEN have the best metric compared to all other 5 validation sets suggesting that the DL framework is biased on degradation model is used for the training. However, the model trained with training data using all 6 degradation models (combined model), but with the same T, has consistent and robust performance across all validation data. The same conclusion can be derived from Figure 4 which shows the average PSNR for all experiments. Figure 5 shows a representative super-resolution image reflecting the bias of a DL framework trained on bicubic modeling of low-resolution images.

Conclusion

We investigated the bias of deep learning frameworks on training data. We found that the degradation model produces poor results when the degradation of the test set differs from that of the training set. Deep learning super-resolution framework trained and tested on incorrectly modeled low-resolution and high-resolution image pairs can give misleading performance metrics.

Acknowledgements

No acknowledgement found.

References

1. Christian Ledig, Lucas Theis, Ferenc Huszar, Jose Caballero, Andrew P. Aitken, Alykhan Tejani,Johannes Totz, Zehan Wang, and Wenzhe Shi. Photo-realistic single image super-resolutionusing a generative adversarial network. CoRR, abs/1609.04802, 2016.

2. R. Keys. Cubic convolution interpolation for digital image processing. IEEE Transactions onAcoustics, Speech, and Signal Processing, 29(6):1153–1160, 1981

3. R.Hamming Digital filters (Mathematics).Englewood Cliffs, NJ : Prentice-Hall,1983.

4. Xintao Wang, Ke Yu, Shixiang Wu, Jinjin Gu, Yihao Liu, Chao Dong, Chen Change Loy,Yu Qiao, and Xiaoou Tang. ESRGAN: enhanced super-resolution generative adversarialnetworks. CoRR, abs/1809.00219, 2018.

5. Akella Ravi Tej, Shirsendu Sukanta Halder, Arunav Pratap Shandeelya, and Vinod Pankajakshan.Enhancing perceptual loss with adversarial feature matching for super-resolution, 2020.

6. Aamir Mustafa, Aliaksei Mikhailiuk, Dan Andrei Iliescu, Varun Babbar, and Rafal K. Mantiuk.Training a task-specific image reconstruction loss, 2021.

7. Gao Huang, Zhuang Liu, and Kilian Q. Weinberger. Densely connected convolutional networks.CoRR, abs/1608.06993, 2016.

8. Saiprasad Ravishankar and Yoram Bresler. Mr image reconstruction from highly undersampledk-space data by dictionary learning. IEEE Transactions on Medical Imaging, 30(5):1028–1041,2011.

Figures

Figure 1: Illustration of super-resolution of low-resolution image to obtain high-resolution image.

Figure 2: DenseNet architecture

Figure 3: NRMSE and HFEN plot calculated for model trained on a) gaussian dataset, b) hanning dataset, c) hamming dataset, d) bicubic dataset, e) mean dataset, f) median dataset, g) combine dataset and evaluated on all 6 simulation test set . Figure on the top is a box plot for NRMSE metric and the figure on bottom is for HFEN. From the figure, we can see that a model trained on a combined dataset with more diversity is performing better across all six degradation test sets. Rest of the model gives poor performance when the test set and training set do not follow the same degradation.

Figure 4: Average PSNR for the different test sets. Each column represents the test set and rows represent the model training set. For each test set, a model trained on a similar training set has the best value represented by a bold letter. We can see that the model trained on a combination of all types of degradation on the last row has consistent results across all test sets.

Figure 5: Example results for super-resolution on images from bicubic simulation test set. Zoom-ins of the red region of the output from each of model is shown in the figure. We can see that only bicubic model and combine model have better PSNR and NRMSE compared to input image. Note that gaussian model represent the model trained on gaussian simulated images and same applies to other model name as well.

Proc. Intl. Soc. Mag. Reson. Med. 31 (2023)
2934
DOI: https://doi.org/10.58530/2023/2934