Prediction with Corrupted Expert Advice
Abstract
We revisit the fundamental problem of prediction with expert advice, in a setting where the environment is benign and generates losses stochastically, but the feedback observed by the learner is subject to a moderate adversarial corruption. We prove that a variant of the classical Multiplicative Weights algorithm with decreasing step sizes achieves constant regret in this setting and performs optimally in a wide range of environments, regardless of the magnitude of the injected corruption. Our results reveal a surprising disparity between the often comparable Follow the Regularized Leader (FTRL) and Online Mirror Descent (OMD) frameworks: we show that for experts in the corrupted stochastic regime, the regret performance of OMD is in fact strictly inferior to that of FTRL.
Cite
Text
Amir et al. "Prediction with Corrupted Expert Advice." Neural Information Processing Systems, 2020.Markdown
[Amir et al. "Prediction with Corrupted Expert Advice." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/amir2020neurips-prediction/)BibTeX
@inproceedings{amir2020neurips-prediction,
title = {{Prediction with Corrupted Expert Advice}},
author = {Amir, Idan and Attias, Idan and Koren, Tomer and Mansour, Yishay and Livni, Roi},
booktitle = {Neural Information Processing Systems},
year = {2020},
url = {https://mlanthology.org/neurips/2020/amir2020neurips-prediction/}
}