Alternating Randomized Block Coordinate Descent
Abstract
Block-coordinate descent algorithms and alternating minimization methods are fundamental optimization algorithms and an important primitive in large-scale optimization and machine learning. While various block-coordinate-descent-type methods have been studied extensively, only alternating minimization – which applies to the setting of only two blocks – is known to have convergence time that scales independently of the least smooth block. A natural question is then: is the setting of two blocks special? We show that the answer is “no” as long as the least smooth block can be optimized exactly – an assumption that is also needed in the setting of alternating minimization. We do so by introducing a novel algorithm AR-BCD, whose convergence time scales independently of the least smooth (possibly non-smooth) block. The basic algorithm generalizes both alternating minimization and randomized block coordinate (gradient) descent, and we also provide its accelerated version – AAR-BCD.
Cite
Text
Diakonikolas and Orecchia. "Alternating Randomized Block Coordinate Descent." International Conference on Machine Learning, 2018.Markdown
[Diakonikolas and Orecchia. "Alternating Randomized Block Coordinate Descent." International Conference on Machine Learning, 2018.](https://mlanthology.org/icml/2018/diakonikolas2018icml-alternating/)BibTeX
@inproceedings{diakonikolas2018icml-alternating,
title = {{Alternating Randomized Block Coordinate Descent}},
author = {Diakonikolas, Jelena and Orecchia, Lorenzo},
booktitle = {International Conference on Machine Learning},
year = {2018},
pages = {1224-1232},
volume = {80},
url = {https://mlanthology.org/icml/2018/diakonikolas2018icml-alternating/}
}