Diverse Prototypical Ensembles Improve Robustness to Subpopulation Shift

Abstract

Subpopulation shift, characterized by a disparity in subpopulation distribution between the training and target datasets, can significantly degrade the performance of machine learning models. Current solutions to subpopulation shift involve modifying empirical risk minimization with re-weighting strategies to improve generalization. This strategy relies on assumptions about the number and nature of subpopulations and annotations on group membership, which are unavailable for many real-world datasets. Instead, we propose using an ensemble of diverse classifiers to adaptively capture risk associated with subpopulations. Given a feature extractor network, we replace its standard linear classification layer with a mixture of prototypical classifiers, where each member is trained to classify the data while focusing on different features and samples from other members. In empirical evaluation on nine real-world datasets, covering diverse domains and kinds of subpopulation shift, our method of Diverse Prototypical Ensembles (DPEs) often outperforms the prior state-of-the-art in worst-group accuracy. The code is available at https://github.com/minhto2802/dpe4subpop.

Cite

Text

To et al. "Diverse Prototypical Ensembles Improve Robustness to Subpopulation Shift." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[To et al. "Diverse Prototypical Ensembles Improve Robustness to Subpopulation Shift." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/to2025icml-diverse/)

BibTeX

@inproceedings{to2025icml-diverse,
  title     = {{Diverse Prototypical Ensembles Improve Robustness to Subpopulation Shift}},
  author    = {To, Nguyen Nhat Minh and Wilson, Paul F R and Nguyen, Viet and Harmanani, Mohamed and Cooper, Michael and Fooladgar, Fahimeh and Abolmaesumi, Purang and Mousavi, Parvin and Krishnan, Rahul},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {59761-59783},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/to2025icml-diverse/}
}