Self-Calibrating Probability Forecasting

Abstract

In the problem of probability forecasting the learner’s goal is to output, given a training set and a new object, a suitable probability measure on the possible values of the new object’s label. An on-line algorithm for probability forecasting is said to be well-calibrated if the probabilities it outputs agree with the observed frequencies. We give a natural non- asymptotic formalization of the notion of well-calibratedness, which we then study under the assumption of randomness (the object/label pairs are independent and identically distributed). It turns out that, although no probability forecasting algorithm is automatically well-calibrated in our sense, there exists a wide class of algorithms for “multiprobability forecasting” (such algorithms are allowed to output a set, ideally very narrow, of probability measures) which satisfy this property; we call the algorithms in this class “Venn probability machines”. Our experimental results demonstrate that a 1-Nearest Neighbor Venn probability machine performs reasonably well on a standard benchmark data set, and one of our theoretical results asserts that a simple Venn probability machine asymptotically approaches the true conditional probabilities regardless, and without knowledge, of the true probability measure generating the examples.

Cite

Text

Vovk et al. "Self-Calibrating Probability Forecasting." Neural Information Processing Systems, 2003.

Markdown

[Vovk et al. "Self-Calibrating Probability Forecasting." Neural Information Processing Systems, 2003.](https://mlanthology.org/neurips/2003/vovk2003neurips-selfcalibrating/)

BibTeX

@inproceedings{vovk2003neurips-selfcalibrating,
  title     = {{Self-Calibrating Probability Forecasting}},
  author    = {Vovk, Vladimir and Shafer, Glenn and Nouretdinov, Ilia},
  booktitle = {Neural Information Processing Systems},
  year      = {2003},
  pages     = {1133-1140},
  url       = {https://mlanthology.org/neurips/2003/vovk2003neurips-selfcalibrating/}
}