Beyond Top-Class Agreement: Using Divergences to Forecast Performance Under Distribution Shift

Abstract

Knowing if a model will generalize to data `in the wild' is crucial for safe deployment. To this end, we study model disagreement notions that consider the full predictive distribution - specifically disagreement based on Hellinger distance, Jensen-Shannon and Kullback–Leibler divergence. We find that divergence-based scores provide better test error estimates and detection rates on out-of-distribution data compared to their top-1 counterparts. Experiments involve standard vision and foundation models.

Cite

Text

Schirmer et al. "Beyond Top-Class Agreement: Using Divergences to Forecast Performance Under Distribution Shift." NeurIPS 2023 Workshops: DistShift, 2023.

Markdown

[Schirmer et al. "Beyond Top-Class Agreement: Using Divergences to Forecast Performance Under Distribution Shift." NeurIPS 2023 Workshops: DistShift, 2023.](https://mlanthology.org/neuripsw/2023/schirmer2023neuripsw-beyond/)

BibTeX

@inproceedings{schirmer2023neuripsw-beyond,
  title     = {{Beyond Top-Class Agreement: Using Divergences to Forecast Performance Under Distribution Shift}},
  author    = {Schirmer, Mona and Zhang, Dan and Nalisnick, Eric},
  booktitle = {NeurIPS 2023 Workshops: DistShift},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/schirmer2023neuripsw-beyond/}
}