Optimal Sampling in Unbiased Active Learning

Abstract

A common belief in unbiased active learning is that, in order to capture the most informative instances, the sampling probabilities should be proportional to the uncertainty of the class labels. We argue that this produces suboptimal predictions and present sampling schemes for unbiased pool-based active learning that minimise the actual prediction error, and demonstrate a better predictive performance than competing methods on a number of benchmark datasets. In contrast, both probabilistic and deterministic uncertainty sampling performed worse than simple random sampling on some of the datasets.

Cite

Text

Imberg et al. "Optimal Sampling in Unbiased Active Learning." Artificial Intelligence and Statistics, 2020.

Markdown

[Imberg et al. "Optimal Sampling in Unbiased Active Learning." Artificial Intelligence and Statistics, 2020.](https://mlanthology.org/aistats/2020/imberg2020aistats-optimal/)

BibTeX

@inproceedings{imberg2020aistats-optimal,
  title     = {{Optimal Sampling in Unbiased Active Learning}},
  author    = {Imberg, Henrik and Jonasson, Johan and Axelson-Fisk, Marina},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2020},
  pages     = {559-569},
  volume    = {108},
  url       = {https://mlanthology.org/aistats/2020/imberg2020aistats-optimal/}
}