Semi-Supervised Learning with Meta-Gradient
Abstract
In this work, we propose a simple yet effective meta-learning algorithm in semi-supervised learning. We notice that most existing consistency-based approaches suffer from overfitting and limited model generalization ability, especially when training with only a small number of labeled data. To alleviate this issue, we propose a learn-to-generalize regularization term by utilizing the label information and optimize the problem in a meta-learning fashion. Specifically, we seek the pseudo labels of the unlabeled data so that the model can generalize well on the labeled data, which is formulated as a nested optimization problem. We address this problem using the meta-gradient that bridges between the pseudo label and the regularization term. In addition, we introduce a simple first-order approximation to avoid computing higher-order derivatives and provide theoretic convergence analysis. Extensive evaluations on the SVHN, CIFAR, and ImageNet datasets demonstrate that the proposed algorithm performs favorably against state-of-the-art methods.
Cite
Text
Xiao et al. "Semi-Supervised Learning with Meta-Gradient." Artificial Intelligence and Statistics, 2021.Markdown
[Xiao et al. "Semi-Supervised Learning with Meta-Gradient." Artificial Intelligence and Statistics, 2021.](https://mlanthology.org/aistats/2021/xiao2021aistats-semisupervised/)BibTeX
@inproceedings{xiao2021aistats-semisupervised,
title = {{Semi-Supervised Learning with Meta-Gradient}},
author = {Xiao, Taihong and Zhang, Xin-Yu and Jia, Haolin and Cheng, Ming-Ming and Yang, Ming-Hsuan},
booktitle = {Artificial Intelligence and Statistics},
year = {2021},
pages = {73-81},
volume = {130},
url = {https://mlanthology.org/aistats/2021/xiao2021aistats-semisupervised/}
}