How Does Semi-Supervised Learning with Pseudo-Labelers Work? a Case Study
Abstract
Semi-supervised learning is a popular machine learning paradigm that utilizes a large amount of unlabeled data as well as a small amount of labeled data to facilitate learning tasks. While semi-supervised learning has achieved great success in training neural networks, its theoretical understanding remains largely open. In this paper, we aim to theoretically understand a semi-supervised learning approach based on pre-training and linear probing. In particular, the semi-supervised learning approach we consider first trains a two-layer neural network based on the unlabeled data with the help of pseudo-labelers. Then it linearly probes the pre-trained network on a small amount of labeled data. We prove that, under a certain toy data generation model and two-layer convolutional neural network, the semisupervised learning approach can achieve nearly zero test loss, while a neural network directly trained by supervised learning on the same amount of labeled data can only achieve constant test loss. Through this case study, we demonstrate a separation between semi-supervised learning and supervised learning in terms of test loss provided the same amount of labeled data.
Cite
Text
Kou et al. "How Does Semi-Supervised Learning with Pseudo-Labelers Work? a Case Study." International Conference on Learning Representations, 2023.Markdown
[Kou et al. "How Does Semi-Supervised Learning with Pseudo-Labelers Work? a Case Study." International Conference on Learning Representations, 2023.](https://mlanthology.org/iclr/2023/kou2023iclr-semisupervised/)BibTeX
@inproceedings{kou2023iclr-semisupervised,
title = {{How Does Semi-Supervised Learning with Pseudo-Labelers Work? a Case Study}},
author = {Kou, Yiwen and Chen, Zixiang and Cao, Yuan and Gu, Quanquan},
booktitle = {International Conference on Learning Representations},
year = {2023},
url = {https://mlanthology.org/iclr/2023/kou2023iclr-semisupervised/}
}