(Un)certainty Selection Methods for Active Learning on Label Distributions

Abstract

Some supervised learning problems can require predicting a probability distribution over more than one possible (set of) answer(s). In such cases, a major scaling issue is the amount of labels needed since, compared to their single- or multi-label counterparts, distributional labels are typically (1) harder to learn and (2) more expensive to obtain for training and testing. In this paper, we explore the use of active learning to alleviate this bottleneck. We progressively train a label distribution learning model by selectively labeling data and, achieving the minimum error rate with fifty percent fewer data items than non-active learning strategies. Our experiments show that certainty-based query strategies outperform uncertainty-based ones on the label distribution learning problems we study.

Cite

Text

Spann et al. "(Un)certainty Selection Methods for Active Learning on Label Distributions." NeurIPS 2023 Workshops: OPT, 2023.

Markdown

[Spann et al. "(Un)certainty Selection Methods for Active Learning on Label Distributions." NeurIPS 2023 Workshops: OPT, 2023.](https://mlanthology.org/neuripsw/2023/spann2023neuripsw-un/)

BibTeX

@inproceedings{spann2023neuripsw-un,
  title     = {{(Un)certainty Selection Methods for Active Learning on Label Distributions}},
  author    = {Spann, James and Bongale, Pratik Sanjay and Homan, Christopher M},
  booktitle = {NeurIPS 2023 Workshops: OPT},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/spann2023neuripsw-un/}
}