Mitigating Confirmation Bias in Semi-Supervised Learning via Efficient Bayesian Model Averaging
Abstract
State-of-the-art (SOTA) semi-supervised learning (SSL) methods have been highly successful in leveraging a mix of labeled and unlabeled data, often via self-training or pseudo-labeling. During pseudo-labeling, the model's predictions on unlabeled data are used for training and may result in confirmation bias where the model reinforces its own mistakes. In this work, we show that SOTA SSL methods often suffer from confirmation bias and demonstrate that this is often a result of using a poorly calibrated classifier for pseudo labeling. We introduce BaM-SSL, an efficient Bayesian Model averaging technique that improves uncertainty quantification in SSL methods with limited computational or memory overhead. We demonstrate that BaM-SSL mitigates confirmation bias in SOTA SSL methods across standard vision benchmarks of CIFAR-10, CIFAR-100, giving up to 16% improvement in test accuracy on the CIFAR-100 with 400 labels benchmark. Furthermore, we also demonstrate their effectiveness in additional realistic and challenging problems, such as class-imbalanced datasets and in photonics science.
Cite
Text
Loh et al. "Mitigating Confirmation Bias in Semi-Supervised Learning via Efficient Bayesian Model Averaging." Transactions on Machine Learning Research, 2023.Markdown
[Loh et al. "Mitigating Confirmation Bias in Semi-Supervised Learning via Efficient Bayesian Model Averaging." Transactions on Machine Learning Research, 2023.](https://mlanthology.org/tmlr/2023/loh2023tmlr-mitigating/)BibTeX
@article{loh2023tmlr-mitigating,
title = {{Mitigating Confirmation Bias in Semi-Supervised Learning via Efficient Bayesian Model Averaging}},
author = {Loh, Charlotte and Dangovski, Rumen and Sudalairaj, Shivchander and Han, Seungwook and Han, Ligong and Karlinsky, Leonid and Soljacic, Marin and Srivastava, Akash},
journal = {Transactions on Machine Learning Research},
year = {2023},
url = {https://mlanthology.org/tmlr/2023/loh2023tmlr-mitigating/}
}