Fast Asynchronous Parallel Stochastic Gradient Descent: A Lock-Free Approach with Convergence Guarantee
Abstract
Stochastic gradient descent (SGD) and its variants have become more and more popular in machine learning due to their efficiency and effectiveness. To handle large-scale problems, researchers have recently proposed several parallel SGD methods for multicore systems. However, existing parallel SGD methods cannot achieve satisfactory performance in real applications. In this paper, we propose a fast asynchronous parallel SGD method, called AsySVRG, by designing an asynchronous strategy to parallelize the recently proposed SGD variant called stochastic variance reduced gradient (SVRG). AsySVRG adopts a lock-free strategy which is more efficient than other strategies with locks. Furthermore, we theoretically prove that AsySVRG is convergent with a linear convergence rate. Both theoretical and empirical results show that AsySVRG can outperform existing state-of-the-art parallel SGD methods like Hogwild! in terms of convergence rate and computation cost.
Cite
Text
Zhao and Li. "Fast Asynchronous Parallel Stochastic Gradient Descent: A Lock-Free Approach with Convergence Guarantee." AAAI Conference on Artificial Intelligence, 2016. doi:10.1609/AAAI.V30I1.10305Markdown
[Zhao and Li. "Fast Asynchronous Parallel Stochastic Gradient Descent: A Lock-Free Approach with Convergence Guarantee." AAAI Conference on Artificial Intelligence, 2016.](https://mlanthology.org/aaai/2016/zhao2016aaai-fast/) doi:10.1609/AAAI.V30I1.10305BibTeX
@inproceedings{zhao2016aaai-fast,
title = {{Fast Asynchronous Parallel Stochastic Gradient Descent: A Lock-Free Approach with Convergence Guarantee}},
author = {Zhao, Shen-Yi and Li, Wu-Jun},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2016},
pages = {2379-2385},
doi = {10.1609/AAAI.V30I1.10305},
url = {https://mlanthology.org/aaai/2016/zhao2016aaai-fast/}
}