Truthfulness of Calibration Measures

Abstract

We study calibration measures in a sequential prediction setup. In addition to rewarding accurate predictions (completeness) and penalizing incorrect ones (soundness), an important desideratum of calibration measures is truthfulness, a minimal condition for the forecaster not to be incentivized to exploit the system. Formally, a calibration measure is truthful if the forecaster (approximately) minimizes the expected penalty by predicting the conditional expectation of the next outcome, given the prior distribution of outcomes. We conduct a taxonomy of existing calibration measures. Perhaps surprisingly, all of them are far from being truthful. We introduce a new calibration measure termed the Subsampled Smooth Calibration Error (SSCE), which is complete and sound, and under which truthful prediction is optimal up to a constant multiplicative factor. In contrast, under existing calibration measures, there are simple distributions on which a polylogarithmic (or even zero) penalty is achievable, while truthful prediction leads to a polynomial penalty.

Cite

Text

Haghtalab et al. "Truthfulness of Calibration Measures." Neural Information Processing Systems, 2024. doi:10.52202/079017-3722

Markdown

[Haghtalab et al. "Truthfulness of Calibration Measures." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/haghtalab2024neurips-truthfulness/) doi:10.52202/079017-3722

BibTeX

@inproceedings{haghtalab2024neurips-truthfulness,
  title     = {{Truthfulness of Calibration Measures}},
  author    = {Haghtalab, Nika and Qiao, Mingda and Yang, Kunhe and Zhao, Eric},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-3722},
  url       = {https://mlanthology.org/neurips/2024/haghtalab2024neurips-truthfulness/}
}