Revisiting the Calibration of Modern Neural Networks
Abstract
Accurate estimation of predictive uncertainty (model calibration) is essential for the safe application of neural networks. Many instances of miscalibration in modern neural networks have been reported, suggesting a trend that newer, more accurate models produce poorly calibrated predictions. Here, we revisit this question for recent state-of-the-art image classification models. We systematically relate model calibration and accuracy, and find that the most recent models, notably those not using convolutions, are among the best calibrated. Trends observed in prior model generations, such as decay of calibration with distribution shift or model size, are less pronounced in recent architectures. We also show that model size and amount of pretraining do not fully explain these differences, suggesting that architecture is a major determinant of calibration properties.
Cite
Text
Minderer et al. "Revisiting the Calibration of Modern Neural Networks." Neural Information Processing Systems, 2021.Markdown
[Minderer et al. "Revisiting the Calibration of Modern Neural Networks." Neural Information Processing Systems, 2021.](https://mlanthology.org/neurips/2021/minderer2021neurips-revisiting/)BibTeX
@inproceedings{minderer2021neurips-revisiting,
title = {{Revisiting the Calibration of Modern Neural Networks}},
author = {Minderer, Matthias and Djolonga, Josip and Romijnders, Rob and Hubis, Frances and Zhai, Xiaohua and Houlsby, Neil and Tran, Dustin and Lucic, Mario},
booktitle = {Neural Information Processing Systems},
year = {2021},
url = {https://mlanthology.org/neurips/2021/minderer2021neurips-revisiting/}
}