Active Learning Under Label Shift

Abstract

We address the problem of active learning under label shift: when the class proportions of source and target domains differ. We introduce a "medial distribution" to incorporate a tradeoff between importance weighting and class-balanced sampling and propose their combined usage in active learning. Our method is known as Mediated Active Learning under Label Shift (MALLS). It balances the bias from class-balanced sampling and the variance from importance weighting. We prove sample complexity and generalization guarantees for MALLS which show active learning reduces asymptotic sample complexity even under arbitrary label shift. We empirically demonstrate MALLS scales to high-dimensional datasets and can reduce the sample complexity of active learning by 60% in deep active learning tasks.

Cite

Text

Zhao et al. "Active Learning Under Label Shift." Artificial Intelligence and Statistics, 2021.

Markdown

[Zhao et al. "Active Learning Under Label Shift." Artificial Intelligence and Statistics, 2021.](https://mlanthology.org/aistats/2021/zhao2021aistats-active/)

BibTeX

@inproceedings{zhao2021aistats-active,
  title     = {{Active Learning Under Label Shift}},
  author    = {Zhao, Eric and Liu, Anqi and Anandkumar, Animashree and Yue, Yisong},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2021},
  pages     = {3412-3420},
  volume    = {130},
  url       = {https://mlanthology.org/aistats/2021/zhao2021aistats-active/}
}