Asynchronous Distributed Semi-Stochastic Gradient Optimization

Abstract

With the recent proliferation of large-scale learning problems, there have been a lot of interest on distributed machine learning algorithms, particularly those that are based on stochastic gradient descent (SGD) and its variants. However, existing algorithms either suffer from slow convergence due to the inherent variance of stochastic gradients, or have a fast linear convergence rate but at the expense of poorer solution quality. In this paper, we combine their merits by proposing a fast distributed asynchronous SGD-based algorithm with variance reduction. A constant learning rate can be used, and it is also guaranteed to converge linearly to the optimal solution. Experiments on the Google Cloud Computing Platform demonstrate that the proposed algorithm outperforms state-of-the-art distributed asynchronous algorithms in terms of both wall clock time and solution quality.

Cite

Text

Zhang et al. "Asynchronous Distributed Semi-Stochastic Gradient Optimization." AAAI Conference on Artificial Intelligence, 2016. doi:10.1609/AAAI.V30I1.10286

Markdown

[Zhang et al. "Asynchronous Distributed Semi-Stochastic Gradient Optimization." AAAI Conference on Artificial Intelligence, 2016.](https://mlanthology.org/aaai/2016/zhang2016aaai-asynchronous/) doi:10.1609/AAAI.V30I1.10286

BibTeX

@inproceedings{zhang2016aaai-asynchronous,
  title     = {{Asynchronous Distributed Semi-Stochastic Gradient Optimization}},
  author    = {Zhang, Ruiliang and Zheng, Shuai and Kwok, James T.},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2016},
  pages     = {2323-2329},
  doi       = {10.1609/AAAI.V30I1.10286},
  url       = {https://mlanthology.org/aaai/2016/zhang2016aaai-asynchronous/}
}