Decentralized Optimization with Edge Sampling
Abstract
In this paper, we propose a decentralized distributed algorithm with stochastic communication among nodes, building on a sampling method called "edge sampling''. Such a sampling algorithm allows us to avoid the heavy peer-to-peer communication cost when combining neighboring weights on dense networks while still maintains a comparable convergence rate. In particular, we quantitatively analyze its theoretical convergence properties, as well as the optimal sampling rate over the underlying network. When compared with previous methods, our solution is shown to be unbiased, communication-efficient and suffers from lower sampling variances. These theoretical findings are validated by both numerical experiments on the mixing rates of Markov Chains and distributed machine learning problems.
Cite
Text
Zhang et al. "Decentralized Optimization with Edge Sampling." International Joint Conference on Artificial Intelligence, 2019. doi:10.24963/IJCAI.2019/93Markdown
[Zhang et al. "Decentralized Optimization with Edge Sampling." International Joint Conference on Artificial Intelligence, 2019.](https://mlanthology.org/ijcai/2019/zhang2019ijcai-decentralized/) doi:10.24963/IJCAI.2019/93BibTeX
@inproceedings{zhang2019ijcai-decentralized,
title = {{Decentralized Optimization with Edge Sampling}},
author = {Zhang, Chi and Li, Qianxiao and Zhao, Peilin},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2019},
pages = {658-664},
doi = {10.24963/IJCAI.2019/93},
url = {https://mlanthology.org/ijcai/2019/zhang2019ijcai-decentralized/}
}