Non-IID Distributed Learning with Optimal Mixture Weights
Abstract
Distributed learning can well solve the problem of training model with large-scale data, which has attracted much attention in recent years. However, most existing distributed learning algorithms set uniform mixture weights across clients when aggregating the global model, which impairs the accuracy under Non-IID (Not Independently or Identically Distributed) setting. In this paper, we present a general framework to optimize the mixture weights and show that our framework has lower expected loss than the uniform mixture weights framework theoretically. Moreover, we provide strong generalization guarantee for our framework, where the excess risk bound can converge at $\mathcal {O}(1/n)$ O ( 1 / n ) , which is as fast as centralized training. Motivated by the theoretical findings, we propose a novel algorithm to improve the performance of distributed learning under Non-IID setting. Through extensive experiments, we show that our algorithm outperforms other mainstream methods, which coincides with our theory.
Cite
Text
Li et al. "Non-IID Distributed Learning with Optimal Mixture Weights." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2022. doi:10.1007/978-3-031-26412-2_33Markdown
[Li et al. "Non-IID Distributed Learning with Optimal Mixture Weights." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2022.](https://mlanthology.org/ecmlpkdd/2022/li2022ecmlpkdd-noniid/) doi:10.1007/978-3-031-26412-2_33BibTeX
@inproceedings{li2022ecmlpkdd-noniid,
title = {{Non-IID Distributed Learning with Optimal Mixture Weights}},
author = {Li, Jian and Wei, Bojian and Liu, Yong and Wang, Weiping},
booktitle = {European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases},
year = {2022},
pages = {539-554},
doi = {10.1007/978-3-031-26412-2_33},
url = {https://mlanthology.org/ecmlpkdd/2022/li2022ecmlpkdd-noniid/}
}