Adaptive Online Estimation of Piecewise Polynomial Trends
Abstract
We consider the framework of non-stationary stochastic optimization [Besbes et.al. 2015] with squared error losses and noisy gradient feedback where the dynamic regret of an online learner against a time varying comparator sequence is studied. Motivated from the theory of non-parametric regression, we introduce a \emph{new variational constraint} that enforces the comparator sequence to belong to a discrete $k^{th}$ order Total Variation ball of radius $C_n$. This variational constraint models comparators that have piece-wise polynomial structure which has many relevant practical applications [Tibshirani2015]. By establishing connections to the theory of wavelet based non-parametric regression, we design a \emph{polynomial time} algorithm that achieves the nearly \emph{optimal dynamic regret} of $\tilde{O}(n^{\frac{1}{2k+3}}C_n^{\frac{2}{2k+3}})$. The proposed policy is \emph{adaptive to the unknown radius} $C_n$. Further, we show that the same policy is minimax optimal for several other non-parametric families of interest.
Cite
Text
Baby and Wang. "Adaptive Online Estimation of Piecewise Polynomial Trends." Neural Information Processing Systems, 2020.Markdown
[Baby and Wang. "Adaptive Online Estimation of Piecewise Polynomial Trends." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/baby2020neurips-adaptive/)BibTeX
@inproceedings{baby2020neurips-adaptive,
title = {{Adaptive Online Estimation of Piecewise Polynomial Trends}},
author = {Baby, Dheeraj and Wang, Yu-Xiang},
booktitle = {Neural Information Processing Systems},
year = {2020},
url = {https://mlanthology.org/neurips/2020/baby2020neurips-adaptive/}
}