Semi-Supervised Learning Under Class Distribution Mismatch

Abstract

Semi-supervised learning (SSL) aims to avoid the need for collecting prohibitively expensive labelled training data. Whilst demonstrating impressive performance boost, existing SSL methods artificially assume that small labelled data and large unlabelled data are drawn from the same class distribution. In a more realistic scenario with class distribution mismatch between the two sets, they often suffer severe performance degradation due to error propagation introduced by irrelevant unlabelled samples. Our work addresses this under-studied and realistic SSL problem by a novel algorithm named Uncertainty-Aware Self-Distillation (UASD). Specifically, UASD produces soft targets that avoid catastrophic error propagation, and empower learning effectively from unconstrained unlabelled data with out-of-distribution (OOD) samples. This is based on joint Self-Distillation and OOD filtering in a unified formulation. Without bells and whistles, UASD significantly outperforms six state-of-the-art methods in more realistic SSL under class distribution mismatch on three popular image classification datasets: CIFAR10, CIFAR100, and TinyImageNet.

Cite

Text

Chen et al. "Semi-Supervised Learning Under Class Distribution Mismatch." AAAI Conference on Artificial Intelligence, 2020. doi:10.1609/AAAI.V34I04.5763

Markdown

[Chen et al. "Semi-Supervised Learning Under Class Distribution Mismatch." AAAI Conference on Artificial Intelligence, 2020.](https://mlanthology.org/aaai/2020/chen2020aaai-semi/) doi:10.1609/AAAI.V34I04.5763

BibTeX

@inproceedings{chen2020aaai-semi,
  title     = {{Semi-Supervised Learning Under Class Distribution Mismatch}},
  author    = {Chen, Yanbei and Zhu, Xiatian and Li, Wei and Gong, Shaogang},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2020},
  pages     = {3569-3576},
  doi       = {10.1609/AAAI.V34I04.5763},
  url       = {https://mlanthology.org/aaai/2020/chen2020aaai-semi/}
}