On Calibration of Modern Neural Networks

Abstract

Confidence calibration – the problem of predicting probability estimates representative of the true correctness likelihood – is important for classification models in many applications. We discover that modern neural networks, unlike those from a decade ago, are poorly calibrated. Through extensive experiments, we observe that depth, width, weight decay, and Batch Normalization are important factors influencing calibration. We evaluate the performance of various post-processing calibration methods on state-of-the-art architectures with image and document classification datasets. Our analysis and experiments not only offer insights into neural network learning, but also provide a simple and straightforward recipe for practical settings: on most datasets, temperature scaling – a single-parameter variant of Platt Scaling – is surprisingly effective at calibrating predictions.

Cite

Text

Guo et al. "On Calibration of Modern Neural Networks." International Conference on Machine Learning, 2017.

Markdown

[Guo et al. "On Calibration of Modern Neural Networks." International Conference on Machine Learning, 2017.](https://mlanthology.org/icml/2017/guo2017icml-calibration/)

BibTeX

@inproceedings{guo2017icml-calibration,
  title     = {{On Calibration of Modern Neural Networks}},
  author    = {Guo, Chuan and Pleiss, Geoff and Sun, Yu and Weinberger, Kilian Q.},
  booktitle = {International Conference on Machine Learning},
  year      = {2017},
  pages     = {1321-1330},
  volume    = {70},
  url       = {https://mlanthology.org/icml/2017/guo2017icml-calibration/}
}