On Calibration of Modern Neural Networks
Abstract
Confidence calibration – the problem of predicting probability estimates representative of the true correctness likelihood – is important for classification models in many applications. We discover that modern neural networks, unlike those from a decade ago, are poorly calibrated. Through extensive experiments, we observe that depth, width, weight decay, and Batch Normalization are important factors influencing calibration. We evaluate the performance of various post-processing calibration methods on state-of-the-art architectures with image and document classification datasets. Our analysis and experiments not only offer insights into neural network learning, but also provide a simple and straightforward recipe for practical settings: on most datasets, temperature scaling – a single-parameter variant of Platt Scaling – is surprisingly effective at calibrating predictions.
Cite
Text
Guo et al. "On Calibration of Modern Neural Networks." International Conference on Machine Learning, 2017.Markdown
[Guo et al. "On Calibration of Modern Neural Networks." International Conference on Machine Learning, 2017.](https://mlanthology.org/icml/2017/guo2017icml-calibration/)BibTeX
@inproceedings{guo2017icml-calibration,
title = {{On Calibration of Modern Neural Networks}},
author = {Guo, Chuan and Pleiss, Geoff and Sun, Yu and Weinberger, Kilian Q.},
booktitle = {International Conference on Machine Learning},
year = {2017},
pages = {1321-1330},
volume = {70},
url = {https://mlanthology.org/icml/2017/guo2017icml-calibration/}
}