Open Problems in Efficient Semi-Supervised PAC Learning

Abstract

The standard PAC model focuses on learning a class of functions from labeled examples, where the two critical resources are the number of examples needed and running time. In many natural learning problems, however, unlabeled data can be obtained much more cheaply than labeled data. This has motivated the notion of semi-supervised learning, in which algorithms attempt to use this cheap unlabeled data in a way that (hopefully) reduces the number of labeled examples needed for learning [4]. For instance, semi-supervised and transductive SVM [2,5] and co-training [3] are two examples of semi-supervised learning algorithms. In [1], a semi-supervised PAC model is introduced that provides a common framework for the kinds of assumptions these algorithms make; however, most of the results in [1] deal with sample complexity rather than computational efficiency, or are only computationally efficient under strong assumptions on the underlying distribution. This note poses several questions related to developing computationally efficient algorithms in this semi-supervised PAC model.

Cite

Text

Blum and Balcan. "Open Problems in Efficient Semi-Supervised PAC Learning." Annual Conference on Computational Learning Theory, 2007. doi:10.1007/978-3-540-72927-3_45

Markdown

[Blum and Balcan. "Open Problems in Efficient Semi-Supervised PAC Learning." Annual Conference on Computational Learning Theory, 2007.](https://mlanthology.org/colt/2007/blum2007colt-open/) doi:10.1007/978-3-540-72927-3_45

BibTeX

@inproceedings{blum2007colt-open,
  title     = {{Open Problems in Efficient Semi-Supervised PAC Learning}},
  author    = {Blum, Avrim and Balcan, Maria-Florina},
  booktitle = {Annual Conference on Computational Learning Theory},
  year      = {2007},
  pages     = {622-624},
  doi       = {10.1007/978-3-540-72927-3_45},
  url       = {https://mlanthology.org/colt/2007/blum2007colt-open/}
}