Generalization to Unseen Cases

Abstract

We analyze classification error on unseen cases, i.e. cases that are different from those in the training set. Unlike standard generalization error, this off-training-set error may differ significantly from the empirical error with high probability even with large sample sizes. We derive a datadependent bound on the difference between off-training-set and standard generalization error. Our result is based on a new bound on the missing mass, which for small samples is stronger than existing bounds based on Good-Turing estimators. As we demonstrate on UCI data-sets, our bound gives nontrivial generalization guarantees in many practical cases. In light of these results, we show that certain claims made in the No Free Lunch literature are overly pessimistic.

Cite

Text

Roos et al. "Generalization to Unseen Cases." Neural Information Processing Systems, 2005.

Markdown

[Roos et al. "Generalization to Unseen Cases." Neural Information Processing Systems, 2005.](https://mlanthology.org/neurips/2005/roos2005neurips-generalization/)

BibTeX

@inproceedings{roos2005neurips-generalization,
  title     = {{Generalization to Unseen Cases}},
  author    = {Roos, Teemu and Grünwald, Peter and Myllymäki, Petri and Tirri, Henry},
  booktitle = {Neural Information Processing Systems},
  year      = {2005},
  pages     = {1129-1136},
  url       = {https://mlanthology.org/neurips/2005/roos2005neurips-generalization/}
}