On the Unreasonable Effectiveness of Federated Averaging with Heterogeneous Data
Abstract
Existing theoretical results (such as (Woodworth et al., 2020a)) predict that the performance of federated averaging (FedAvg) is exacerbated by high data heterogeneity. However, in practice, FedAvg converges pretty well on several naturally heterogeneous datasets. In order to explain this seemingly unreasonable effectiveness of FedAvg that contradicts previous theoretical predictions, this paper introduces the client consensus hypothesis: on certain federated datasets, the average of local models updates on clients starting from the optimum is close to zero. We prove that under this hypothesis, data heterogeneity does not exacerbate the convergence of FedAvg. Moreover, we show that this hypothesis holds for a linear regression problem and some naturally heterogeneous datasets such as FEMNIST and StackOverflow. Therefore, we believe that this hypothesis can better explain the performance of FedAvg in practice.
Cite
Text
Wang et al. "On the Unreasonable Effectiveness of Federated Averaging with Heterogeneous Data." Transactions on Machine Learning Research, 2024.Markdown
[Wang et al. "On the Unreasonable Effectiveness of Federated Averaging with Heterogeneous Data." Transactions on Machine Learning Research, 2024.](https://mlanthology.org/tmlr/2024/wang2024tmlr-unreasonable-a/)BibTeX
@article{wang2024tmlr-unreasonable-a,
title = {{On the Unreasonable Effectiveness of Federated Averaging with Heterogeneous Data}},
author = {Wang, Jianyu and Das, Rudrajit and Joshi, Gauri and Kale, Satyen and Xu, Zheng and Zhang, Tong},
journal = {Transactions on Machine Learning Research},
year = {2024},
url = {https://mlanthology.org/tmlr/2024/wang2024tmlr-unreasonable-a/}
}