Stateful ODE-Nets Using Basis Function Expansions

Abstract

The recently-introduced class of ordinary differential equation networks (ODE-Nets) establishes a fruitful connection between deep learning and dynamical systems. In this work, we reconsider formulations of the weights as continuous-in-depth functions using linear combinations of basis functions which enables us to leverage parameter transformations such as function projections. In turn, this view allows us to formulate a novel stateful ODE-Block that handles stateful layers. The benefits of this new ODE-Block are twofold: first, it enables incorporating meaningful continuous-in-depth batch normalization layers to achieve state-of-the-art performance; second, it enables compressing the weights through a change of basis, without retraining, while maintaining near state-of-the-art performance and reducing both inference time and memory footprint. Performance is demonstrated by applying our stateful ODE-Block to (a) image classification tasks using convolutional units and (b) sentence-tagging tasks using transformer encoder units.

Cite

Text

Queiruga et al. "Stateful ODE-Nets Using Basis Function Expansions." Neural Information Processing Systems, 2021.

Markdown

[Queiruga et al. "Stateful ODE-Nets Using Basis Function Expansions." Neural Information Processing Systems, 2021.](https://mlanthology.org/neurips/2021/queiruga2021neurips-stateful/)

BibTeX

@inproceedings{queiruga2021neurips-stateful,
  title     = {{Stateful ODE-Nets Using Basis Function Expansions}},
  author    = {Queiruga, Alejandro and Erichson, N. Benjamin and Hodgkinson, Liam and Mahoney, Michael W.},
  booktitle = {Neural Information Processing Systems},
  year      = {2021},
  url       = {https://mlanthology.org/neurips/2021/queiruga2021neurips-stateful/}
}