Nystrom Regularization for Time Series Forecasting

Abstract

This paper focuses on learning rate analysis of Nystrom regularization with sequential sub-sampling for $\tau$-mixing time series. Using a recently developed Banach-valued Bernstein inequality for $\tau$-mixing sequences and an integral operator approach based on second-order decomposition, we succeed in deriving almost optimal learning rates of Nystrom regularization with sequential sub-sampling for $\tau$-mixing time series. A series of numerical experiments are carried out to verify our theoretical results, showing the excellent learning performance of Nystrom regularization with sequential sub-sampling in learning massive time series data. All these results extend the applicable range of Nystr\"om regularization from i.i.d. samples to non-i.i.d. sequences.

Cite

Text

Sun et al. "Nystrom Regularization for Time Series Forecasting." Journal of Machine Learning Research, 2022.

Markdown

[Sun et al. "Nystrom Regularization for Time Series Forecasting." Journal of Machine Learning Research, 2022.](https://mlanthology.org/jmlr/2022/sun2022jmlr-nystrom/)

BibTeX

@article{sun2022jmlr-nystrom,
  title     = {{Nystrom Regularization for Time Series Forecasting}},
  author    = {Sun, Zirui and Dai, Mingwei and Wang, Yao and Lin, Shao-Bo},
  journal   = {Journal of Machine Learning Research},
  year      = {2022},
  pages     = {1-42},
  volume    = {23},
  url       = {https://mlanthology.org/jmlr/2022/sun2022jmlr-nystrom/}
}