Optimizing Importance Weighting in the Presence of Sub-Population Shifts

Abstract

A distribution shift between the training and test data can severely harm performance of machine learning models. Importance weighting addresses this issue by assigning different weights to data points during training. We argue that existing heuristics for determining the weights are suboptimal, as they neglect the increase of the variance of the estimated model due to the limited sample size of the training data. We interpret the optimal weights in terms of a bias-variance trade-off, and propose a bi-level optimization procedure in which the weights and model parameters are optimized simultaneously. We apply this framework to existing importance weighting techniques for last-layer retraining of deep neural networks in the presence of sub-population shifts and show empirically that optimizing weights significantly improves generalization performance.

Cite

Text

Holstege et al. "Optimizing Importance Weighting in the Presence of Sub-Population Shifts." International Conference on Learning Representations, 2025.

Markdown

[Holstege et al. "Optimizing Importance Weighting in the Presence of Sub-Population Shifts." International Conference on Learning Representations, 2025.](https://mlanthology.org/iclr/2025/holstege2025iclr-optimizing/)

BibTeX

@inproceedings{holstege2025iclr-optimizing,
  title     = {{Optimizing Importance Weighting in the Presence of Sub-Population Shifts}},
  author    = {Holstege, Floris and Wouters, Bram and Van Giersbergen, Noud and Diks, Cees},
  booktitle = {International Conference on Learning Representations},
  year      = {2025},
  url       = {https://mlanthology.org/iclr/2025/holstege2025iclr-optimizing/}
}