Trading Off Mistakes and Don't-Know Predictions

Abstract

We discuss an online learning framework in which the agent is allowed to say ``I don't know'' as well as making incorrect predictions on given examples. We analyze the trade off between saying ``I don't know'' and making mistakes. If the number of don't know predictions is forced to be zero, the model reduces to the well-known mistake-bound model introduced by Littlestone [Lit88]. On the other hand, if no mistakes are allowed, the model reduces to KWIK framework introduced by Li et. al. [LLW08]. We propose a general, though inefficient, algorithm for general finite concept classes that minimizes the number of don't-know predictions if a certain number of mistakes are allowed. We then present specific polynomial-time algorithms for the concept classes of monotone disjunctions and linear separators.

Cite

Text

Sayedi et al. "Trading Off Mistakes and Don't-Know Predictions." Neural Information Processing Systems, 2010.

Markdown

[Sayedi et al. "Trading Off Mistakes and Don't-Know Predictions." Neural Information Processing Systems, 2010.](https://mlanthology.org/neurips/2010/sayedi2010neurips-trading/)

BibTeX

@inproceedings{sayedi2010neurips-trading,
  title     = {{Trading Off Mistakes and Don't-Know Predictions}},
  author    = {Sayedi, Amin and Zadimoghaddam, Morteza and Blum, Avrim},
  booktitle = {Neural Information Processing Systems},
  year      = {2010},
  pages     = {2092-2100},
  url       = {https://mlanthology.org/neurips/2010/sayedi2010neurips-trading/}
}