Using Pre-Training Can Improve Model Robustness and Uncertainty

Abstract

He et al. (2018) have called into question the utility of pre-training by showing that training from scratch can often yield similar performance to pre-training. We show that although pre-training may not improve performance on traditional classification metrics, it improves model robustness and uncertainty estimates. Through extensive experiments on label corruption, class imbalance, adversarial examples, out-of-distribution detection, and confidence calibration, we demonstrate large gains from pre-training and complementary effects with task-specific methods. We show approximately a 10% absolute improvement over the previous state-of-the-art in adversarial robustness. In some cases, using pre-training without task-specific methods also surpasses the state-of-the-art, highlighting the need for pre-training when evaluating future methods on robustness and uncertainty tasks.

Cite

Text

Hendrycks et al. "Using Pre-Training Can Improve Model Robustness and Uncertainty." International Conference on Machine Learning, 2019.

Markdown

[Hendrycks et al. "Using Pre-Training Can Improve Model Robustness and Uncertainty." International Conference on Machine Learning, 2019.](https://mlanthology.org/icml/2019/hendrycks2019icml-using/)

BibTeX

@inproceedings{hendrycks2019icml-using,
  title     = {{Using Pre-Training Can Improve Model Robustness and Uncertainty}},
  author    = {Hendrycks, Dan and Lee, Kimin and Mazeika, Mantas},
  booktitle = {International Conference on Machine Learning},
  year      = {2019},
  pages     = {2712-2721},
  volume    = {97},
  url       = {https://mlanthology.org/icml/2019/hendrycks2019icml-using/}
}