A Randomized Online Learning Algorithm for Better Variance Control
Abstract
We propose a sequential randomized algorithm, which at each step concentrates on functions having both low risk and low variance with respect to the previous step prediction function. It satisfies a simple risk bound, which is sharp to the extent that the standard statistical learning approach, based on supremum of empirical processes, does not lead to algorithms with such a tight guarantee on its efficiency. Our generalization error bounds complement the pioneering work of Cesa-Bianchi et al. [12] in which standard-style statistical results were recovered with tight constants using worst-case analysis. A nice feature of our analysis of the randomized estimator is to put forward the links between the probabilistic and worst-case viewpoint. It also allows to recover recent model selection results due to Juditsky et al. [16] and to improve them in least square regression with heavy noise, i.e. when no exponential moment condition is assumed on the output.
Cite
Text
Audibert. "A Randomized Online Learning Algorithm for Better Variance Control." Annual Conference on Computational Learning Theory, 2006. doi:10.1007/11776420_30Markdown
[Audibert. "A Randomized Online Learning Algorithm for Better Variance Control." Annual Conference on Computational Learning Theory, 2006.](https://mlanthology.org/colt/2006/audibert2006colt-randomized/) doi:10.1007/11776420_30BibTeX
@inproceedings{audibert2006colt-randomized,
title = {{A Randomized Online Learning Algorithm for Better Variance Control}},
author = {Audibert, Jean-Yves},
booktitle = {Annual Conference on Computational Learning Theory},
year = {2006},
pages = {392-407},
doi = {10.1007/11776420_30},
url = {https://mlanthology.org/colt/2006/audibert2006colt-randomized/}
}