Fast Asynchronous Parallel Stochastic Gradient Descent: A Lock-Free Approach with Convergence Guarantee

Abstract

Stochastic gradient descent (SGD) and its variants have become more and more popular in machine learning due to their efficiency and effectiveness. To handle large-scale problems, researchers have recently proposed several parallel SGD methods for multicore systems. However, existing parallel SGD methods cannot achieve satisfactory performance in real applications. In this paper, we propose a fast asynchronous parallel SGD method, called AsySVRG, by designing an asynchronous strategy to parallelize the recently proposed SGD variant called stochastic variance reduced gradient (SVRG). AsySVRG adopts a lock-free strategy which is more efficient than other strategies with locks. Furthermore, we theoretically prove that AsySVRG is convergent with a linear convergence rate. Both theoretical and empirical results show that AsySVRG can outperform existing state-of-the-art parallel SGD methods like Hogwild! in terms of convergence rate and computation cost.

Cite

Text

Zhao and Li. "Fast Asynchronous Parallel Stochastic Gradient Descent: A Lock-Free Approach with Convergence Guarantee." AAAI Conference on Artificial Intelligence, 2016. doi:10.1609/AAAI.V30I1.10305

Markdown

[Zhao and Li. "Fast Asynchronous Parallel Stochastic Gradient Descent: A Lock-Free Approach with Convergence Guarantee." AAAI Conference on Artificial Intelligence, 2016.](https://mlanthology.org/aaai/2016/zhao2016aaai-fast/) doi:10.1609/AAAI.V30I1.10305

BibTeX

@inproceedings{zhao2016aaai-fast,
  title     = {{Fast Asynchronous Parallel Stochastic Gradient Descent: A Lock-Free Approach with Convergence Guarantee}},
  author    = {Zhao, Shen-Yi and Li, Wu-Jun},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2016},
  pages     = {2379-2385},
  doi       = {10.1609/AAAI.V30I1.10305},
  url       = {https://mlanthology.org/aaai/2016/zhao2016aaai-fast/}
}