Scaling up Semi-Supervised Learning with Unconstrained Unlabelled Data

Abstract

We propose UnMixMatch, a semi-supervised learning framework which can learn effective representations from unconstrained unlabelled data in order to scale up performance. Most existing semi-supervised methods rely on the assumption that labelled and unlabelled samples are drawn from the same distribution, which limits the potential for improvement through the use of free-living unlabeled data. Consequently, the generalizability and scalability of semi-supervised learning are often hindered by this assumption. Our method aims to overcome these constraints and effectively utilize unconstrained unlabelled data in semi-supervised learning. UnMixMatch consists of three main components: a supervised learner with hard augmentations that provides strong regularization, a contrastive consistency regularizer to learn underlying representations from the unlabelled data, and a self-supervised loss to enhance the representations that are learnt from the unlabelled data. We perform extensive experiments on 4 commonly used datasets and demonstrate superior performance over existing semi-supervised methods with a performance boost of 4.79%. Extensive ablation and sensitivity studies show the effectiveness and impact of each of the proposed components of our method. The code for our work is publicly available.

Cite

Text

Roy and Etemad. "Scaling up Semi-Supervised Learning with Unconstrained Unlabelled Data." AAAI Conference on Artificial Intelligence, 2024. doi:10.1609/AAAI.V38I13.29404

Markdown

[Roy and Etemad. "Scaling up Semi-Supervised Learning with Unconstrained Unlabelled Data." AAAI Conference on Artificial Intelligence, 2024.](https://mlanthology.org/aaai/2024/roy2024aaai-scaling/) doi:10.1609/AAAI.V38I13.29404

BibTeX

@inproceedings{roy2024aaai-scaling,
  title     = {{Scaling up Semi-Supervised Learning with Unconstrained Unlabelled Data}},
  author    = {Roy, Shuvendu and Etemad, Ali},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2024},
  pages     = {14847-14856},
  doi       = {10.1609/AAAI.V38I13.29404},
  url       = {https://mlanthology.org/aaai/2024/roy2024aaai-scaling/}
}