Top-Ambiguity Samples Matter: Understanding Why Deep Ensemble Works in Selective Classification

Abstract

Selective classification allows a machine learning model to reject some hard inputs and thus improve the reliability of its predictions. In this area, the ensemble method is powerful in practice, but there has been no solid analysis on why the ensemble method works. Inspired by an interesting empirical result that the improvement of the ensemble largely comes from top-ambiguity samples where its member models diverge, we prove that, based on some assumptions, the ensemble has a lower selective risk than the member model for any coverage within a range. The proof is nontrivial since the selective risk is a non-convex function of the model prediction. The assumptions and the theoretical results are supported by systematic experiments on both computer vision and natural language processing tasks.

Cite

Text

Ding et al. "Top-Ambiguity Samples Matter: Understanding Why Deep Ensemble Works in Selective Classification." Neural Information Processing Systems, 2023.

Markdown

[Ding et al. "Top-Ambiguity Samples Matter: Understanding Why Deep Ensemble Works in Selective Classification." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/ding2023neurips-topambiguity/)

BibTeX

@inproceedings{ding2023neurips-topambiguity,
  title     = {{Top-Ambiguity Samples Matter: Understanding Why Deep Ensemble Works in Selective Classification}},
  author    = {Ding, Qiang and Cao, Yixuan and Luo, Ping},
  booktitle = {Neural Information Processing Systems},
  year      = {2023},
  url       = {https://mlanthology.org/neurips/2023/ding2023neurips-topambiguity/}
}