FedMut: Generalized Federated Learning via Stochastic Mutation

Abstract

Although Federated Learning (FL) enables collaborative model training without sharing the raw data of clients, it encounters low-performance problems caused by various heterogeneous scenarios. Due to the limitation of dispatching the same global model to clients for local training, traditional Federated Average (FedAvg)-based FL models face the problem of easily getting stuck into a sharp solution, which results in training a low-performance global model. To address this problem, this paper presents a novel FL approach named FedMut, which mutates the global model according to the gradient change to generate several intermediate models for the next round of training. Each intermediate model will be dispatched to a client for local training. Eventually, the global model converges into a flat area within the range of mutated models and has a well-generalization compared with the global model trained by FedAvg. Experimental results on well-known datasets demonstrate the effectiveness of our FedMut approach in various data heterogeneity scenarios.

Cite

Text

Hu et al. "FedMut: Generalized Federated Learning via Stochastic Mutation." AAAI Conference on Artificial Intelligence, 2024. doi:10.1609/AAAI.V38I11.29146

Markdown

[Hu et al. "FedMut: Generalized Federated Learning via Stochastic Mutation." AAAI Conference on Artificial Intelligence, 2024.](https://mlanthology.org/aaai/2024/hu2024aaai-fedmut/) doi:10.1609/AAAI.V38I11.29146

BibTeX

@inproceedings{hu2024aaai-fedmut,
  title     = {{FedMut: Generalized Federated Learning via Stochastic Mutation}},
  author    = {Hu, Ming and Cao, Yue and Li, Anran and Li, Zhiming and Liu, Chengwei and Li, Tianlin and Chen, Mingsong and Liu, Yang},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2024},
  pages     = {12528-12537},
  doi       = {10.1609/AAAI.V38I11.29146},
  url       = {https://mlanthology.org/aaai/2024/hu2024aaai-fedmut/}
}