Communication-Efficient Heterogeneous Federated Learning with Generalized Heavy-Ball Momentum
Abstract
Federated Learning (FL) has emerged as the state-of-the-art approach for learning from decentralized data in privacy-constrained scenarios. However, system and statistical challenges hinder its real-world applicability, requiring efficient learning from edge devices and robustness to data heterogeneity. Despite significant research efforts, existing approaches often degrade severely due to the joint effect of heterogeneity and partial client participation. In particular, while momentum appears as a promising approach for overcoming statistical heterogeneity, in current approaches its update is biased towards the most recently sampled clients. As we show in this work, this is the reason why it fails to outperform FedAvg, preventing its effective use in real-world large-scale scenarios. In this work, we propose a novel Generalized Heavy-Ball Momentum (GHBM) and theoretically prove it enables convergence under unbounded data heterogeneity in cyclic partial participation, thereby advancing the understanding of momentum's effectiveness in FL. We then introduce adaptive and communication-efficient variants of GHBM that match the communication complexity of FedAvg in settings where clients can be stateful. Extensive experiments on vision and language tasks confirm our theoretical findings, demonstrating that GHBM substantially improves state-of-the-art performance under random uniform client sampling, particularly in large-scale settings with high data heterogeneity and low client participation.
Cite
Text
Zaccone et al. "Communication-Efficient Heterogeneous Federated Learning with Generalized Heavy-Ball Momentum." Transactions on Machine Learning Research, 2025.Markdown
[Zaccone et al. "Communication-Efficient Heterogeneous Federated Learning with Generalized Heavy-Ball Momentum." Transactions on Machine Learning Research, 2025.](https://mlanthology.org/tmlr/2025/zaccone2025tmlr-communicationefficient/)BibTeX
@article{zaccone2025tmlr-communicationefficient,
title = {{Communication-Efficient Heterogeneous Federated Learning with Generalized Heavy-Ball Momentum}},
author = {Zaccone, Riccardo and Karimireddy, Sai Praneeth and Masone, Carlo and Ciccone, Marco},
journal = {Transactions on Machine Learning Research},
year = {2025},
url = {https://mlanthology.org/tmlr/2025/zaccone2025tmlr-communicationefficient/}
}