Communication/Computation Tradeoffs in Consensus-Based Distributed Optimization
Abstract
We study the scalability of consensus-based distributed optimization algorithms by considering two questions: How many processors should we use for a given problem, and how often should they communicate when communication is not free? Central to our analysis is a problem-specific value $r$ which quantifies the communication/computation tradeoff. We show that organizing the communication among nodes as a $k$-regular expander graph~\cite{kRegExpanders} yields speedups, while when all pairs of nodes communicate (as in a complete graph), there is an optimal number of processors that depends on $r$. Surprisingly, a speedup can be obtained, in terms of the time to reach a fixed level of accuracy, by communicating less and less frequently as the computation progresses. Experiments on a real cluster solving metric learning and non-smooth convex minimization tasks demonstrate strong agreement between theory and practice.
Cite
Text
Tsianos et al. "Communication/Computation Tradeoffs in Consensus-Based Distributed Optimization." Neural Information Processing Systems, 2012.Markdown
[Tsianos et al. "Communication/Computation Tradeoffs in Consensus-Based Distributed Optimization." Neural Information Processing Systems, 2012.](https://mlanthology.org/neurips/2012/tsianos2012neurips-communication/)BibTeX
@inproceedings{tsianos2012neurips-communication,
title = {{Communication/Computation Tradeoffs in Consensus-Based Distributed Optimization}},
author = {Tsianos, Konstantinos and Lawlor, Sean and Rabbat, Michael G.},
booktitle = {Neural Information Processing Systems},
year = {2012},
pages = {1943-1951},
url = {https://mlanthology.org/neurips/2012/tsianos2012neurips-communication/}
}