Probabilistic Federated Prompt-Tuning with Non-IID and Imbalanced Data

Abstract

Fine-tuning pre-trained models is a popular approach in machine learning for solving complex tasks with moderate data. However, fine-tuning the entire pre-trained model is ineffective in federated data scenarios where local data distributions are diversely skewed. To address this, we explore integrating federated learning with a more effective prompt-tuning method, optimizing for a small set of input prefixes to reprogram the pre-trained model's behavior. Our approach transforms federated learning into a distributed set modeling task, aggregating diverse sets of prompts to globally fine-tune the pre-trained model. We benchmark various baselines based on direct adaptations of existing federated model aggregation techniques and introduce a new probabilistic prompt aggregation method that substantially outperforms these baselines. Our reported results on a variety of computer vision datasets confirm that the proposed method is most effective to combat extreme data heterogeneity in federated learning.

Cite

Text

Weng et al. "Probabilistic Federated Prompt-Tuning with Non-IID and Imbalanced Data." Neural Information Processing Systems, 2024. doi:10.52202/079017-2604

Markdown

[Weng et al. "Probabilistic Federated Prompt-Tuning with Non-IID and Imbalanced Data." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/weng2024neurips-probabilistic/) doi:10.52202/079017-2604

BibTeX

@inproceedings{weng2024neurips-probabilistic,
  title     = {{Probabilistic Federated Prompt-Tuning with Non-IID and Imbalanced Data}},
  author    = {Weng, Pei-Yau and Hoang, Minh and Nguyen, Lam M. and Thai, My T. and Weng, Tsui-Wei and Hoang, Trong Nghia},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-2604},
  url       = {https://mlanthology.org/neurips/2024/weng2024neurips-probabilistic/}
}