An Empirical Study of Federated Prompt Learning for Vision Language Model

Abstract

The Vision Language Model (VLM) excels in aligning vision and language representations, and prompt learning has emerged as a key technique for adapting such models to downstream tasks. However, the application of prompt learning with VLM in federated learning (FL) scenarios remains underexplored. This paper systematically investigates the behavioral differences between language prompt learning (LPT) and vision prompt learning (VPT) under data heterogeneity challenges, including label skew and domain shift. We conduct extensive experiments to evaluate the impact of various FL and prompt configurations, such as client scale, aggregation strategies, and prompt length, to assess the robustness of Federated Prompt Learning (FPL). Furthermore, we explore strategies for enhancing prompt learning in complex scenarios where label skew and domain shift coexist, including leveraging both prompt types when computational resources allow. Our findings offer practical insights into optimizing prompt learning in federated settings, contributing to the broader deployment of VLMs in privacy-preserving environments.

Cite

Text

Wang et al. "An Empirical Study of Federated Prompt Learning for Vision Language Model." International Joint Conference on Artificial Intelligence, 2025. doi:10.24963/IJCAI.2025/1188

Markdown

[Wang et al. "An Empirical Study of Federated Prompt Learning for Vision Language Model." International Joint Conference on Artificial Intelligence, 2025.](https://mlanthology.org/ijcai/2025/wang2025ijcai-empirical/) doi:10.24963/IJCAI.2025/1188

BibTeX

@inproceedings{wang2025ijcai-empirical,
  title     = {{An Empirical Study of Federated Prompt Learning for Vision Language Model}},
  author    = {Wang, Zhihao and Huang, Wenke and Chen, Tian and Shi, Zekun and Wan, Guancheng and Qiao, Yu and Yang, Bin and Wang, Jian and Li, Bing and Ye, Mang},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {10705-10713},
  doi       = {10.24963/IJCAI.2025/1188},
  url       = {https://mlanthology.org/ijcai/2025/wang2025ijcai-empirical/}
}