Enhancing Foundation Models with Federated Domain Knowledge Infusion
Abstract
Vision foundation models (FMs) like CLIP have exhibited exceptional capabilities in visual and linguistic understanding, particularly in zero-shot inference tasks. However, these models struggle with data that significantly deviates from their training samples, necessitating fine-tuning, which is often infeasible in centralized settings due to data privacy concerns. Federated learning (FL) combined with parameter-efficient fine-tuning (PEFT) offers a potential solution, yet existing methods face issues with domain-specific characteristics and out-of-domain generalization. We propose a cross-silo Federated Adapter Generalization (FedAG), a novel federated fine-tuning approach that leverages multiple fine-grained adapters to capture domain-specific knowledge while enhancing out-of-domain generalization. Our method uses quality-aware in-domain mutual learning and attention-regularized cross-domain learning to integrate domain-specific insights effectively. Experiments of the CLIP model on three domain-shifting datasets, ImageCLEF-DA, Office-Home, and DomainNet, demonstrate the superior performance of FedAG in both in-domain and out-of-domain scenarios. We envision this work as a milestone for generalizing CLIP to handle the challenge of out-of-domain knowledge under federated learning setting.
Cite
Text
Wang et al. "Enhancing Foundation Models with Federated Domain Knowledge Infusion." Proceedings of the 42nd International Conference on Machine Learning, 2025.Markdown
[Wang et al. "Enhancing Foundation Models with Federated Domain Knowledge Infusion." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/wang2025icml-enhancing/)BibTeX
@inproceedings{wang2025icml-enhancing,
title = {{Enhancing Foundation Models with Federated Domain Knowledge Infusion}},
author = {Wang, Jiaqi and Li, Jingtao and Zhuang, Weiming and Chen, Chen and Lyu, Lingjuan and Ma, Fenglong},
booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
year = {2025},
pages = {63621-63635},
volume = {267},
url = {https://mlanthology.org/icml/2025/wang2025icml-enhancing/}
}