Semi-ViM: Bidirectional State Space Model for Mitigating Label Imbalance in Semi-Supervised Learning
Abstract
Semi-supervised learning (SSL) is often hindered by learning biases when imbalanced datasets are used for training, which limits its effectiveness in real-world applications. In this paper, we propose Semi-ViM, a novel SSL framework based on Vision Mamba, a bidirectional state space model (SSM) that serves as a superior alternative to Transformer-based architectures for visual representation learning. Semi-ViM effectively deals with label imbalance and improves model stability through two key innovations: LyapEMA, a stability-aware parameter update mechanism inspired by Lyapunov theory, and SSMixup, a novel mixup strategy applied at the hidden state level of bidirectional SSMs. Experimental results on ImageNet-1K and ImageNet-LT demonstrate that Semi-ViM significantly outperforms state-of-the-art SSL models, achieving 85.40% accuracy with only 10% of the labeled data, surpassing Transformer-based methods such as Semi-ViT.
Cite
Text
He et al. "Semi-ViM: Bidirectional State Space Model for Mitigating Label Imbalance in Semi-Supervised Learning." International Conference on Computer Vision, 2025.Markdown
[He et al. "Semi-ViM: Bidirectional State Space Model for Mitigating Label Imbalance in Semi-Supervised Learning." International Conference on Computer Vision, 2025.](https://mlanthology.org/iccv/2025/he2025iccv-semivim/)BibTeX
@inproceedings{he2025iccv-semivim,
title = {{Semi-ViM: Bidirectional State Space Model for Mitigating Label Imbalance in Semi-Supervised Learning}},
author = {He, Hongyang and Xie, Hongyang and You, Haochen and Sanchez, Victor},
booktitle = {International Conference on Computer Vision},
year = {2025},
pages = {765-774},
url = {https://mlanthology.org/iccv/2025/he2025iccv-semivim/}
}