Semi-Supervised Learning with Meta-Gradient

Abstract

In this work, we propose a simple yet effective meta-learning algorithm in semi-supervised learning. We notice that most existing consistency-based approaches suffer from overfitting and limited model generalization ability, especially when training with only a small number of labeled data. To alleviate this issue, we propose a learn-to-generalize regularization term by utilizing the label information and optimize the problem in a meta-learning fashion. Specifically, we seek the pseudo labels of the unlabeled data so that the model can generalize well on the labeled data, which is formulated as a nested optimization problem. We address this problem using the meta-gradient that bridges between the pseudo label and the regularization term. In addition, we introduce a simple first-order approximation to avoid computing higher-order derivatives and provide theoretic convergence analysis. Extensive evaluations on the SVHN, CIFAR, and ImageNet datasets demonstrate that the proposed algorithm performs favorably against state-of-the-art methods.

Cite

Text

Xiao et al. "Semi-Supervised Learning with Meta-Gradient." Artificial Intelligence and Statistics, 2021.

Markdown

[Xiao et al. "Semi-Supervised Learning with Meta-Gradient." Artificial Intelligence and Statistics, 2021.](https://mlanthology.org/aistats/2021/xiao2021aistats-semisupervised/)

BibTeX

@inproceedings{xiao2021aistats-semisupervised,
  title     = {{Semi-Supervised Learning with Meta-Gradient}},
  author    = {Xiao, Taihong and Zhang, Xin-Yu and Jia, Haolin and Cheng, Ming-Ming and Yang, Ming-Hsuan},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2021},
  pages     = {73-81},
  volume    = {130},
  url       = {https://mlanthology.org/aistats/2021/xiao2021aistats-semisupervised/}
}