Towards Making Unlabeled Data Never Hurt
Abstract
It is usually expected that, when labeled data are limited, the learning performance can be improved by exploiting unlabeled data. In many cases, however, the performances of current semi-supervised learning approaches may be even worse than purely using the limited labeled data. It is desired to have \textit{safe} semi-supervised learning approaches which never degenerate learning performance by using unlabeled data. In this paper, we focus on semi-supervised support vector machines (S3VMs) and propose S4VMs, i.e., safe S3VMs. Unlike S3VMs which typically aim at approaching an optimal low-density separator, S4VMs try to exploit the candidate low-density separators simultaneously to reduce the risk of identifying a poor separator with unlabeled data. We describe two implementations of S4VMs, and our comprehensive experiments show that the overall performance of S4VMs are highly competitive to S3VMs, while in contrast to S3VMs which degenerate performance in many cases, S4VMs are never significantly inferior to inductive SVMs.
Cite
Text
Li and Zhou. "Towards Making Unlabeled Data Never Hurt." International Conference on Machine Learning, 2011. doi:10.1109/TPAMI.2014.2299812Markdown
[Li and Zhou. "Towards Making Unlabeled Data Never Hurt." International Conference on Machine Learning, 2011.](https://mlanthology.org/icml/2011/li2011icml-making/) doi:10.1109/TPAMI.2014.2299812BibTeX
@inproceedings{li2011icml-making,
title = {{Towards Making Unlabeled Data Never Hurt}},
author = {Li, Yufeng and Zhou, Zhi-Hua},
booktitle = {International Conference on Machine Learning},
year = {2011},
pages = {1081-1088},
doi = {10.1109/TPAMI.2014.2299812},
url = {https://mlanthology.org/icml/2011/li2011icml-making/}
}