Byzantine Stochastic Gradient Descent

Abstract

This paper studies the problem of distributed stochastic optimization in an adversarial setting where, out of $m$ machines which allegedly compute stochastic gradients every iteration, an $\alpha$-fraction are Byzantine, and may behave adversarially. Our main result is a variant of stochastic gradient descent (SGD) which finds $\varepsilon$-approximate minimizers of convex functions in $T = \tilde{O}\big( \frac{1}{\varepsilon^2 m} + \frac{\alpha^2}{\varepsilon^2} \big)$ iterations. In contrast, traditional mini-batch SGD needs $T = O\big( \frac{1}{\varepsilon^2 m} \big)$ iterations, but cannot tolerate Byzantine failures. Further, we provide a lower bound showing that, up to logarithmic factors, our algorithm is information-theoretically optimal both in terms of sample complexity and time complexity.

Cite

Text

Alistarh et al. "Byzantine Stochastic Gradient Descent." Neural Information Processing Systems, 2018.

Markdown

[Alistarh et al. "Byzantine Stochastic Gradient Descent." Neural Information Processing Systems, 2018.](https://mlanthology.org/neurips/2018/alistarh2018neurips-byzantine/)

BibTeX

@inproceedings{alistarh2018neurips-byzantine,
  title     = {{Byzantine Stochastic Gradient Descent}},
  author    = {Alistarh, Dan and Allen-Zhu, Zeyuan and Li, Jerry},
  booktitle = {Neural Information Processing Systems},
  year      = {2018},
  pages     = {4613-4623},
  url       = {https://mlanthology.org/neurips/2018/alistarh2018neurips-byzantine/}
}