SLowcalSGD : Slow Query Points Improve Local-SGD for Stochastic Convex Optimization
Abstract
We consider distributed learning scenarios where $M$ machines interact with a parameter server along several communication rounds in order to minimize a joint objective function. Focusing on the heterogeneous case, where different machines may draw samples from different data-distributions, we design the first local update method that provably benefits over the two most prominent distributed baselines: namely Minibatch-SGD and Local-SGD. Key to our approach is a slow querying technique that we customize to the distributed setting, which in turn enables a better mitigation of the bias caused by local updates.
Cite
Text
Dahan and Levy. "SLowcalSGD : Slow Query Points Improve Local-SGD for Stochastic Convex Optimization." Neural Information Processing Systems, 2024. doi:10.52202/079017-2957Markdown
[Dahan and Levy. "SLowcalSGD : Slow Query Points Improve Local-SGD for Stochastic Convex Optimization." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/dahan2024neurips-slowcalsgd/) doi:10.52202/079017-2957BibTeX
@inproceedings{dahan2024neurips-slowcalsgd,
title = {{SLowcalSGD : Slow Query Points Improve Local-SGD for Stochastic Convex Optimization}},
author = {Dahan, Tehila and Levy, Kfir Y.},
booktitle = {Neural Information Processing Systems},
year = {2024},
doi = {10.52202/079017-2957},
url = {https://mlanthology.org/neurips/2024/dahan2024neurips-slowcalsgd/}
}