Online Learning with Abstention

Abstract

We present an extensive study of a key problem in online learning where the learner can opt to abstain from making a prediction, at a certain cost. In the adversarial setting, we show how existing online algorithms and guarantees can be adapted to this problem. In the stochastic setting, we first point out a bias problem that limits the straightforward extension of algorithms such as UCB-N to this context. Next, we give a new algorithm, UCB-GT, that exploits historical data and time-varying feedback graphs. We show that this algorithm benefits from more favorable regret guarantees than a natural extension of UCB-N . We further report the results of a series of experiments demonstrating that UCB-GT largely outperforms that extension of UCB-N, as well as other standard baselines.

Cite

Text

Cortes et al. "Online Learning with Abstention." International Conference on Machine Learning, 2018.

Markdown

[Cortes et al. "Online Learning with Abstention." International Conference on Machine Learning, 2018.](https://mlanthology.org/icml/2018/cortes2018icml-online/)

BibTeX

@inproceedings{cortes2018icml-online,
  title     = {{Online Learning with Abstention}},
  author    = {Cortes, Corinna and DeSalvo, Giulia and Gentile, Claudio and Mohri, Mehryar and Yang, Scott},
  booktitle = {International Conference on Machine Learning},
  year      = {2018},
  pages     = {1059-1067},
  volume    = {80},
  url       = {https://mlanthology.org/icml/2018/cortes2018icml-online/}
}