Consistency Matters: Neural ODE Parameters Are Dependent on the Training Numerical Method

Abstract

Neural Ordinary Differential Equations (Neural ODEs) are continuous-depth models that use an ordinary differential equation (ODE) to capture the dynamics of data. Due to their modelling capabilities several works on applications and novel architectures using Neural ODEs can be found in the literature. In this work, we call for the attention to the need of using the same numerical method for both training and making predictions with Neural ODEs since the numerical method employed influences the prediction process, thereby impacting the loss function and introducing variance into parameter optimisation. We provide theoretical insights into how numerical methods of varying orders or with different step sizes influence the loss function of the network. To validate our theoretical analysis, we conduct a series of simple preliminary numerical experiments employing a regression task, demonstrating how the training numerical method influences model performance for testing. Our findings underscore the need for consistency in numerical methods for training and prediction, a consideration not previously emphasised or documented in the literature.

Cite

Text

Coelho et al. "Consistency Matters: Neural ODE Parameters Are Dependent on the Training Numerical Method." ICLR 2024 Workshops: AI4DiffEqtnsInSci, 2024.

Markdown

[Coelho et al. "Consistency Matters: Neural ODE Parameters Are Dependent on the Training Numerical Method." ICLR 2024 Workshops: AI4DiffEqtnsInSci, 2024.](https://mlanthology.org/iclrw/2024/coelho2024iclrw-consistency/)

BibTeX

@inproceedings{coelho2024iclrw-consistency,
  title     = {{Consistency Matters: Neural ODE Parameters Are Dependent on the Training Numerical Method}},
  author    = {Coelho, C. and Costa, M.Fernanda P. and Ferrás, Luís L.},
  booktitle = {ICLR 2024 Workshops: AI4DiffEqtnsInSci},
  year      = {2024},
  url       = {https://mlanthology.org/iclrw/2024/coelho2024iclrw-consistency/}
}