Taming the Wild: A Unified Analysis of Hogwild-Style Algorithms

Abstract

Stochastic gradient descent (SGD) is a ubiquitous algorithm for a variety of machine learning problems. Researchers and industry have developed several techniques to optimize SGD's runtime performance, including asynchronous execution and reduced precision. Our main result is a martingale-based analysis that enables us to capture the rich noise models that may arise from such techniques. Specifically, we useour new analysis in three ways: (1) we derive convergence rates for the convex case (Hogwild) with relaxed assumptions on the sparsity of the problem; (2) we analyze asynchronous SGD algorithms for non-convex matrix problems including matrix completion; and (3) we design and analyze an asynchronous SGD algorithm, called Buckwild, that uses lower-precision arithmetic. We show experimentally that our algorithms run efficiently for a variety of problems on modern hardware.

Cite

Text

De Sa et al. "Taming the Wild: A Unified Analysis of Hogwild-Style Algorithms." Neural Information Processing Systems, 2015.

Markdown

[De Sa et al. "Taming the Wild: A Unified Analysis of Hogwild-Style Algorithms." Neural Information Processing Systems, 2015.](https://mlanthology.org/neurips/2015/sa2015neurips-taming/)

BibTeX

@inproceedings{sa2015neurips-taming,
  title     = {{Taming the Wild: A Unified Analysis of Hogwild-Style Algorithms}},
  author    = {De Sa, Christopher M and Zhang, Ce and Olukotun, Kunle and Ré, Christopher and Ré, Christopher},
  booktitle = {Neural Information Processing Systems},
  year      = {2015},
  pages     = {2674-2682},
  url       = {https://mlanthology.org/neurips/2015/sa2015neurips-taming/}
}