A Self-Training Approach to Cost Sensitive Uncertainty Sampling
Abstract
Uncertainty sampling is an effective method for performing active learning that is computationally efficient compared to other active learning methods such as loss-reduction methods. However, unlike loss-reduction methods, uncertainty sampling cannot minimize total misclassification costs when errors incur different costs. This paper introduces a method for performing cost-sensitive uncertainty sampling that makes use of self-training. We show that, even when misclassification costs are equal, this self-training approach results in faster reduction of loss as a function of number of points labeled and more reliable posterior probability estimates as compared to standard uncertainty sampling. We also show why other more naive methods of modifying uncertainty sampling to minimize total misclassification costs will not always work well.
Cite
Text
Liu et al. "A Self-Training Approach to Cost Sensitive Uncertainty Sampling." Machine Learning, 2009. doi:10.1007/S10994-009-5131-9Markdown
[Liu et al. "A Self-Training Approach to Cost Sensitive Uncertainty Sampling." Machine Learning, 2009.](https://mlanthology.org/mlj/2009/liu2009mlj-selftraining/) doi:10.1007/S10994-009-5131-9BibTeX
@article{liu2009mlj-selftraining,
title = {{A Self-Training Approach to Cost Sensitive Uncertainty Sampling}},
author = {Liu, Alexander and Jun, Goo and Ghosh, Joydeep},
journal = {Machine Learning},
year = {2009},
pages = {257-270},
doi = {10.1007/S10994-009-5131-9},
volume = {76},
url = {https://mlanthology.org/mlj/2009/liu2009mlj-selftraining/}
}