Coordinate Descent with Bandit Sampling

Abstract

Coordinate descent methods minimize a cost function by updating a single decision variable (corresponding to one coordinate) at a time. Ideally, we would update the decision variable that yields the largest marginal decrease in the cost function. However, finding this coordinate would require checking all of them, which is not computationally practical. Therefore, we propose a new adaptive method for coordinate descent. First, we define a lower bound on the decrease of the cost function when a coordinate is updated and, instead of calculating this lower bound for all coordinates, we use a multi-armed bandit algorithm to learn which coordinates result in the largest marginal decrease and simultaneously perform coordinate descent. We show that our approach improves the convergence of the coordinate methods both theoretically and experimentally.

Cite

Text

Salehi et al. "Coordinate Descent with Bandit Sampling." Neural Information Processing Systems, 2018.

Markdown

[Salehi et al. "Coordinate Descent with Bandit Sampling." Neural Information Processing Systems, 2018.](https://mlanthology.org/neurips/2018/salehi2018neurips-coordinate/)

BibTeX

@inproceedings{salehi2018neurips-coordinate,
  title     = {{Coordinate Descent with Bandit Sampling}},
  author    = {Salehi, Farnood and Thiran, Patrick and Celis, Elisa},
  booktitle = {Neural Information Processing Systems},
  year      = {2018},
  pages     = {9247-9257},
  url       = {https://mlanthology.org/neurips/2018/salehi2018neurips-coordinate/}
}