Communication Trade-Offs for Local-SGD with Large Step Size
Abstract
Synchronous mini-batch SGD is state-of-the-art for large-scale distributed machine learning. However, in practice, its convergence is bottlenecked by slow communication rounds between worker nodes. A natural solution to reduce communication is to use the \emph{``local-SGD''} model in which the workers train their model independently and synchronize every once in a while. This algorithm improves the computation-communication trade-off but its convergence is not understood very well. We propose a non-asymptotic error analysis, which enables comparison to \emph{one-shot averaging} i.e., a single communication round among independent workers, and \emph{mini-batch averaging} i.e., communicating at every step. We also provide adaptive lower bounds on the communication frequency for large step-sizes ($ t^{-\alpha} $, $ \alpha\in (1/2 , 1 ) $) and show that \emph{Local-SGD} reduces communication by a factor of $O\Big(\frac{\sqrt{T}}{P^{3/2}}\Big)$, with $T$ the total number of gradients and $P$ machines.
Cite
Text
Dieuleveut and Patel. "Communication Trade-Offs for Local-SGD with Large Step Size." Neural Information Processing Systems, 2019.Markdown
[Dieuleveut and Patel. "Communication Trade-Offs for Local-SGD with Large Step Size." Neural Information Processing Systems, 2019.](https://mlanthology.org/neurips/2019/dieuleveut2019neurips-communication/)BibTeX
@inproceedings{dieuleveut2019neurips-communication,
title = {{Communication Trade-Offs for Local-SGD with Large Step Size}},
author = {Dieuleveut, Aymeric and Patel, Kumar Kshitij},
booktitle = {Neural Information Processing Systems},
year = {2019},
pages = {13601-13612},
url = {https://mlanthology.org/neurips/2019/dieuleveut2019neurips-communication/}
}