0051

The More the Merrier? - On the Number of Trainable Parameters in Iterative Neural Networks for Image Reconstruction
Andreas Kofler1, Tobias Schaeffter1,2,3, and Christoph Kolbitsch1,3
1Physikalisch-Technische Bundesanstalt, Braunschweig and Berlin, Germany, 2Department of Biomedical Engineering, Technische Universität Berlin, Berlin, Germany, 3School of Imaging Sciences and Biomedical Engineering, King's College London, London, United Kingdom

Synopsis

Iterative neural networks (INNs) currently define the state-of-the-art for image reconstruction methods. With these methods, the obtained regularizers are not only optimally adapted to the employed physical model but also tailored to the reconstruction method the network implicitly defines. However, comparing the performance of different INNs-based methods is often challenging because of the black-box character of neural networks. In this work we construct an example which highlights the importance of keeping the number of trainable parameters approximately fixed when comparing INNs-based methods. If this aspect is not taken into consideration, wrong conclusions could be drawn.

Introduction and Motivation

Iterative Neural Networks (INNs) nowadays define the state-of-the-art for image reconstruction problems across all imaging modalities. Their success can be attributed to the fact that the physical model of the considered inverse problem is used in the network architecture. Using INNs, one can learn regularizers which are tailored to the considered forward model and to the employed reconstruction algorithm the network implicitly defines.
Because NNs are still black-boxes from a theoretical point of view, the reason for the superiority of one specific INN compared to others is often unclear. Training with the physical model reduces the maximum-error bound [1], stabilizes the training process [2] and reduces the amount of required training data [3]. Thus, it is distinctly harder to overfit INNs compared to model-agnostic NNs and, to some extent, one can in principle always obtain "better" reconstruction networks by only increasing the number of trainable parameters (NTP) as long as no overfitting is observed. We believe that this aspect - if not taken into consideration - can possibly hinder the development of genuinely novel reconstruction methods because the success of specific methods might mistakenly be attributed to the network-capacity rather than the methods itself.
Here, we consider three INNs-based approaches and demonstrate that by comparing them without indicating NTP, wrong conclusions could be drawn.

Methods

We consider an accelerated 2D radial MR image reconstruction problem
$$\mathbf{A} \mathbf{x} + \mathbf{e} = \mathbf{y},$$
where $$$\mathbf{A}$$$ denotes the encoding operator, $$$\mathbf{x}$$$ the MR image, $$$\mathbf{e}$$$ random Gaussian noise and $$$\mathbf{y}$$$ the undersampled $$$k$$$-space data. In non-Cartesian acquistions, pre-conditioning is often used by weighting the $$$k$$$-space data with a diagonal density-compensation operator $$$\mathbf{W}$$$. The data-discrepancy term and the respective gradient are given by
$$
D_{\mathbf{W}}(\mathbf{A}\mathbf{x},\mathbf{y}) := \frac{1}{2}\|\mathbf{W}^{1/2} (\mathbf{A}\mathbf{x}-\mathbf{y})\|_2^2,\\
\nabla_{\mathbf{x}} D_{\mathbf{W}}(\mathbf{A}\mathbf{x},\mathbf{y}) = \mathbf{A}^{\sharp}(\mathbf{A}\mathbf{x}-\mathbf{y}),
$$
with $$$\mathbf{A}^{\sharp}:=\mathbf{A}^H \mathbf{W}^{1/2}$$$.
We consider three different INNs-based methods based on a learned CNN-projected gradient descent (CNN-Proj) [4], on learned regularizers using a Landweber iteration (CNN-LW) and on a fixed-point iteration using the MoDL-architecture (MoDL) [5]. Note that all three methods can be derived from an underlying functional which defines the reconstruction problem.
Given $$$\mathbf{x}_0:=\mathbf{A}^{\sharp} \mathbf{y}$$$ and an appropriate CNN-architecture $$$f_{\Theta}$$$, for $$$0\leq k\leq T$$$, the methods are given by
$$
\mathbf{z}_k = \mathbf{x}_k - \lambda \, \mathbf{A}^{\sharp}(\mathbf{A}\mathbf{x}-\mathbf{y}), \\ \mathbf{x}_{k+1} = f_{\Theta}(\mathbf{z}_k) \tag{CNN-Proj},
$$
$$
\mathbf{x}_{k+1} = \mathbf{x}_k - \lambda \, f_{\Theta}(\mathbf{x}_{k}) - \lambda \, \mathbf{A}^{\sharp}(\mathbf{A}\mathbf{x}-\mathbf{y}), \tag{CNN-LW}
$$
and
$$
\mathbf{z}_k = f_{\Theta}(\mathbf{x}_k), \\
\mathbf{x}_{k+1} = \underset{\mathbf{x}}{\arg \min} \, \frac{1}{2} \|\mathbf{W}^{1/2} (\mathbf{A}\mathbf{x}-\mathbf{y})\|_2^2 + \frac{\lambda}{2} \, \|\mathbf{x} - \mathbf{z}_k\|_2^2, \tag{MoDL}
$$
where $$$\lambda>0$$$ for MoDL and $$$ 0 < \lambda <2/ \|\mathbf{A}\|^2$$$ for CNN-LW and CNN-Proj. For the three methods, $$$f_{\Theta}$$$ allows for a different interpretation. In CNN-Proj, we can identify $$$f_{\Theta}$$$ to perform a learned projection. In CNN-LW, $$$f_{\Theta}$$$ corresponds to the gradient of a regularizer parametrized by $$$\Theta$$$ and in MoDL, $$$f_{\Theta}$$$ yields a CNN-based image-prior.
Further, note that CNN-Proj and CNN-LW consist of the exact same components, namely the gradient $$$ \nabla_{\mathbf{x}} D_{\mathbf{W}} $$$ and $$$f_{\Theta}$$$. However, CNN-LW is by design more data-consistent compared to CNN-Proj because the last operation is not given by the application of the CNN. Further, MoDL is more data-consistent than CNN-LW, because instead of applying the gradient $$$\nabla_{\mathbf{x}} D_{\mathbf{W}}$$$, it solves a quadratic problem with respect to $$$\mathbf{x}$$$.
By construction, we would therefore expect MoDL to surpass both CNN-LW and CNN-Proj and CNN-LW to surpass CNN-Proj in terms of performance.

For $$$f_{\Theta}$$$, we used a 2D U-Net [6] which we parametrized by $$$E,C$$$ and $$$K$$$, denoting the number of encoding stages, the number of convolutional layers per stage and the initial number of filters, respectively. We trained the three methods with $$$T=6$$$ for different combinations of $$$E=1,...,5 $$$ with $$$C=2$$$ and $$$K=16$$$, yielding different NTPs.
As dataset, we used 2000 images from the MRI dataset in [7], which we split into 1400/400/200 images for training, validation and testing, respectively. The initial reconstructions were retrospectively simulated using an acceleration factor of $$$R\approx 18$$$ with $$$N_c=12$$$ coils. All networks were trained for 500 epochs with ADAM with a learning rate of $$$10^{-4}$$$ using the $$$L_2$$$-error as a loss function.

Results

In Figure 1, we see the training- and validation-errors for the different network architectures we trained on. The superiority of MoDL for this type of imaging problem is clearly visible compared to CNN-LW and CNN-Proj. Note that this observation holds for all different NTP. However, it is also possible to surpass MoDL with CNN-LW and CNN-Proj if NTP is high enough. The box-plots in Figure 2 and the images in Figure 3 highlight the described phenomenon well.

Discussion and Conclusion

Although MoDL consistently outperforms CNN-Proj and CNN-LW at equal network-capacity, increasing NTP makes it possible to surpass MoDL both with CNN-Proj and CNN-LW.
This simple example shows that comparing different INNs without indicating the NTP can indeed yield to a misinterpretation of the performance of the methods. Further, using larger GPUs, one can typically also increase NTPs which also possibly introduces a hardware-bias.
We therefore encourage researchers to always report NTP in their methods and to be aware of the possibility of this issue. Further, we think that NTP could and should be used as a standard metric for evaluating machine learning-based methods for image reconstruction.

Acknowledgements

No acknowledgement found.

References

[1] Maier, Andreas K., et al. "Learning with known operators reduces maximum error bounds." Nature machine intelligence 1.8 (2019): 373-380.

[2] Kofler, Andreas, et al. "An end‐to‐end‐trainable iterative network architecture for accelerated radial multi‐coil 2D cine MR image reconstruction." Medical Physics 48.5 (2021): 2412-2425.

[3] Gilton, Davis, et al. "Neumann networks for linear inverse problems in imaging." IEEE Transactions on Computational Imaging 6 (2019): 328-343.

[4] Gupta, Harshit, et al. "CNN-based projected gradient descent for consistent CT image reconstruction." IEEE transactions on medical imaging 37.6 (2018): 1440-1453.

[5] Aggarwal, Hemant K., et al. "MoDL: Model-based deep learning architecture for inverse problems." IEEE transactions on medical imaging 38.2 (2018): 394-405.

[6] Ronneberger, Olaf, et al. "U-net: Convolutional networks for biomedical image segmentation." International Conference on Medical image computing and computer-assisted intervention. Springer, Cham, 2015.

[7] Bien, Nicholas, et al. "Deep-learning-assisted diagnosis for knee magnetic resonance imaging: development and retrospective validation of MRNet." PLoS medicine 15.11 (2018): e1002699.

Figures

Figure 1: Training- and validation-errors for CNN-LW, CNN-Proj and MoDL for different networks with different network-capacities. All networks generalize well for the first four configurations of NTP. For NTP=2.156,864, overfitting starts to be visible as the validation error is simliar or greater than for NTP=534.848.

Although for each network capacity MoDL exhibits a lower validation-error compared to CNN-LW and CNN-Proj, it is possible for CNN-Proj and CNN-LW to yield similar results compared to MoDL by increasing NTP (e.g. CNN-LW for NTP=129.344 vs MoDL for NTP=2.624).


Figure 2: Results and point-wise error-images for MoDL, CNN-LW and CNN-Proj for different NTP. As can be seen, MoDL yields the smallest point-wise errors compared to CNN-LW and CNN-Proj at the same network-capacity (i.e. same NTP). However, increasing NTP makes it possible for CNN-LW and CNN-Proj to yield a smaller reconstruction-error compared to MoDL with smaller NTP.

Figure 3: Box-plots for the peak signal-to-noise ratio (PSNR), the normalized root mean squared error (NRMSE) and the structural similarity index measure (SSIM) for MoDL, CNN-LW and CNN-Proj for different NTP. Again, the quality of the reconstruction is the best for MoDL for fixed network-capacity with respect to all measures. However, since with increasing NTP (until NTP=534.848) all networks tend to improve the reconstructions with respect to all measures, both CNN-Proj and CNN-LW are able to surpass MoDL for larger NTP.

Proc. Intl. Soc. Mag. Reson. Med. 30 (2022)
0051
DOI: https://doi.org/10.58530/2022/0051