Taylorformer: Probabalistic Modelling for Random Processes Including Time Series
Abstract
We propose the Taylorformer for random processes such as time series. Its two key components are: 1) the LocalTaylor wrapper which adapts Taylor approximations (used in dynamical systems) for use in neural network-based probabilistic models, and 2) the MHA-X attention block which makes predictions in a way inspired by how Gaussian Processes' mean predictions are linear smoothings of contextual data. Taylorformer outperforms the state-of-the-art in terms of log-likelihood on 5/6 classic Neural Process tasks such as meta-learning 1D functions, and has at least a 14\% MSE improvement on forecasting tasks, including electricity, oil temperatures and exchange rates. Taylorformer approximates a consistent stochastic process and provides uncertainty-aware predictions. Our code is provided in the supplementary material.
Cite
Text
Nivron et al. "Taylorformer: Probabalistic Modelling for Random Processes Including Time Series." ICML 2023 Workshops: Frontiers4LCD, 2023.Markdown
[Nivron et al. "Taylorformer: Probabalistic Modelling for Random Processes Including Time Series." ICML 2023 Workshops: Frontiers4LCD, 2023.](https://mlanthology.org/icmlw/2023/nivron2023icmlw-taylorformer/)BibTeX
@inproceedings{nivron2023icmlw-taylorformer,
title = {{Taylorformer: Probabalistic Modelling for Random Processes Including Time Series}},
author = {Nivron, Omer and Parthipan, Raghul and Wischik, Damon},
booktitle = {ICML 2023 Workshops: Frontiers4LCD},
year = {2023},
url = {https://mlanthology.org/icmlw/2023/nivron2023icmlw-taylorformer/}
}