Taylorformer: Probabalistic Modelling for Random Processes Including Time Series

Abstract

We propose the Taylorformer for random processes such as time series. Its two key components are: 1) the LocalTaylor wrapper which adapts Taylor approximations (used in dynamical systems) for use in neural network-based probabilistic models, and 2) the MHA-X attention block which makes predictions in a way inspired by how Gaussian Processes' mean predictions are linear smoothings of contextual data. Taylorformer outperforms the state-of-the-art in terms of log-likelihood on 5/6 classic Neural Process tasks such as meta-learning 1D functions, and has at least a 14\% MSE improvement on forecasting tasks, including electricity, oil temperatures and exchange rates. Taylorformer approximates a consistent stochastic process and provides uncertainty-aware predictions. Our code is provided in the supplementary material.

Cite

Text

Nivron et al. "Taylorformer: Probabalistic Modelling for Random Processes Including Time Series." ICML 2023 Workshops: Frontiers4LCD, 2023.

Markdown

[Nivron et al. "Taylorformer: Probabalistic Modelling for Random Processes Including Time Series." ICML 2023 Workshops: Frontiers4LCD, 2023.](https://mlanthology.org/icmlw/2023/nivron2023icmlw-taylorformer/)

BibTeX

@inproceedings{nivron2023icmlw-taylorformer,
  title     = {{Taylorformer: Probabalistic Modelling for Random Processes Including Time Series}},
  author    = {Nivron, Omer and Parthipan, Raghul and Wischik, Damon},
  booktitle = {ICML 2023 Workshops: Frontiers4LCD},
  year      = {2023},
  url       = {https://mlanthology.org/icmlw/2023/nivron2023icmlw-taylorformer/}
}