The Sample Complexity of Semi-Supervised Learning with Nonparametric Mixture Models

Abstract

We study the sample complexity of semi-supervised learning (SSL) and introduce new assumptions based on the mismatch between a mixture model learned from unlabeled data and the true mixture model induced by the (unknown) class conditional distributions. Under these assumptions, we establish an $\Omega(K\log K)$ labeled sample complexity bound without imposing parametric assumptions, where $K$ is the number of classes. Our results suggest that even in nonparametric settings it is possible to learn a near-optimal classifier using only a few labeled samples. Unlike previous theoretical work which focuses on binary classification, we consider general multiclass classification ($K>2$), which requires solving a difficult permutation learning problem. This permutation defines a classifier whose classification error is controlled by the Wasserstein distance between mixing measures, and we provide finite-sample results characterizing the behaviour of the excess risk of this classifier. Finally, we describe three algorithms for computing these estimators based on a connection to bipartite graph matching, and perform experiments to illustrate the superiority of the MLE over the majority vote estimator.

Cite

Text

Dan et al. "The Sample Complexity of Semi-Supervised Learning with Nonparametric Mixture Models." Neural Information Processing Systems, 2018.

Markdown

[Dan et al. "The Sample Complexity of Semi-Supervised Learning with Nonparametric Mixture Models." Neural Information Processing Systems, 2018.](https://mlanthology.org/neurips/2018/dan2018neurips-sample/)

BibTeX

@inproceedings{dan2018neurips-sample,
  title     = {{The Sample Complexity of Semi-Supervised Learning with Nonparametric Mixture Models}},
  author    = {Dan, Chen and Leqi, Liu and Aragam, Bryon and Ravikumar, Pradeep K and Xing, Eric P},
  booktitle = {Neural Information Processing Systems},
  year      = {2018},
  pages     = {9321-9332},
  url       = {https://mlanthology.org/neurips/2018/dan2018neurips-sample/}
}