Selective Labeling via Error Bound Minimization

Abstract

In many practical machine learning problems, the acquisition of labeled data is often expensive and/or time consuming. This motivates us to study a problem as follows: given a label budget, how to select data points to label such that the learning performance is optimized. We propose a selective labeling method by analyzing the generalization error of Laplacian regularized Least Squares (LapRLS). In particular, we derive a deterministic generalization error bound for LapRLS trained on subsampled data, and propose to select a subset of data points to label by minimizing this upper bound. Since the minimization is a combinational problem, we relax it into continuous domain and solve it by projected gradient descent. Experiments on benchmark datasets show that the proposed method outperforms the state-of-the-art methods.

Cite

Text

Gu et al. "Selective Labeling via Error Bound Minimization." Neural Information Processing Systems, 2012.

Markdown

[Gu et al. "Selective Labeling via Error Bound Minimization." Neural Information Processing Systems, 2012.](https://mlanthology.org/neurips/2012/gu2012neurips-selective/)

BibTeX

@inproceedings{gu2012neurips-selective,
  title     = {{Selective Labeling via Error Bound Minimization}},
  author    = {Gu, Quanquan and Zhang, Tong and Han, Jiawei and Ding, Chris H.},
  booktitle = {Neural Information Processing Systems},
  year      = {2012},
  pages     = {323-331},
  url       = {https://mlanthology.org/neurips/2012/gu2012neurips-selective/}
}