Asynchronous Distributed ADMM for Consensus Optimization

Abstract

Distributed optimization algorithms are highly attractive for solving big data problems. In particular, many machine learning problems can be formulated as the global consensus optimization problem, which can then be solved in a distributed manner by the alternating direction method of multipliers (ADMM) algorithm. However, this suffers from the straggler problem as its updates have to be synchronized. In this paper, we propose an asynchronous ADMM algorithm by using two conditions to control the asynchrony: partial barrier and bounded delay. The proposed algorithm has a simple structure and good convergence guarantees (its convergence rate can be reduced to that of its synchronous counterpart). Experiments on different distributed ADMM applications show that asynchrony reduces the time on network waiting, and achieves faster convergence than its synchronous counterpart in terms of the wall clock time.

Cite

Text

Zhang and Kwok. "Asynchronous Distributed ADMM for Consensus Optimization." International Conference on Machine Learning, 2014.

Markdown

[Zhang and Kwok. "Asynchronous Distributed ADMM for Consensus Optimization." International Conference on Machine Learning, 2014.](https://mlanthology.org/icml/2014/zhang2014icml-asynchronous/)

BibTeX

@inproceedings{zhang2014icml-asynchronous,
  title     = {{Asynchronous Distributed ADMM for Consensus Optimization}},
  author    = {Zhang, Ruiliang and Kwok, James},
  booktitle = {International Conference on Machine Learning},
  year      = {2014},
  pages     = {1701-1709},
  volume    = {32},
  url       = {https://mlanthology.org/icml/2014/zhang2014icml-asynchronous/}
}