Selective Labeling via Error Bound Minimization
Abstract
In many practical machine learning problems, the acquisition of labeled data is often expensive and/or time consuming. This motivates us to study a problem as follows: given a label budget, how to select data points to label such that the learning performance is optimized. We propose a selective labeling method by analyzing the generalization error of Laplacian regularized Least Squares (LapRLS). In particular, we derive a deterministic generalization error bound for LapRLS trained on subsampled data, and propose to select a subset of data points to label by minimizing this upper bound. Since the minimization is a combinational problem, we relax it into continuous domain and solve it by projected gradient descent. Experiments on benchmark datasets show that the proposed method outperforms the state-of-the-art methods.
Cite
Text
Gu et al. "Selective Labeling via Error Bound Minimization." Neural Information Processing Systems, 2012.Markdown
[Gu et al. "Selective Labeling via Error Bound Minimization." Neural Information Processing Systems, 2012.](https://mlanthology.org/neurips/2012/gu2012neurips-selective/)BibTeX
@inproceedings{gu2012neurips-selective,
title = {{Selective Labeling via Error Bound Minimization}},
author = {Gu, Quanquan and Zhang, Tong and Han, Jiawei and Ding, Chris H.},
booktitle = {Neural Information Processing Systems},
year = {2012},
pages = {323-331},
url = {https://mlanthology.org/neurips/2012/gu2012neurips-selective/}
}