Fast Stochastic Composite Minimization and an Accelerated Frank-Wolfe Algorithm Under Parallelization

Abstract

We consider the problem of minimizing the sum of two convex functions. One of those functions has Lipschitz-continuous gradients, and can be accessed via stochastic oracles, whereas the other is ``simple''. We provide a Bregman-type algorithm with accelerated convergence in function values to a ball containing the minimum. The radius of this ball depends on problem-dependent constants, including the variance of the stochastic oracle. We further show that this algorithmic setup naturally leads to a variant of Frank-Wolfe achieving acceleration under parallelization. More precisely, when minimizing a smooth convex function on a bounded domain, we show that one can achieve an $\epsilon$ primal-dual gap (in expectation) in $\tilde{O}(1 /\sqrt{\epsilon})$ iterations, by only accessing gradients of the original function and a linear maximization oracle with $O(1 / \sqrt{\epsilon})$ computing units in parallel. We illustrate this fast convergence on synthetic numerical experiments.

Cite

Text

Dubois-Taine et al. "Fast Stochastic Composite Minimization and an Accelerated Frank-Wolfe Algorithm Under Parallelization." Neural Information Processing Systems, 2022.

Markdown

[Dubois-Taine et al. "Fast Stochastic Composite Minimization and an Accelerated Frank-Wolfe Algorithm Under Parallelization." Neural Information Processing Systems, 2022.](https://mlanthology.org/neurips/2022/duboistaine2022neurips-fast/)

BibTeX

@inproceedings{duboistaine2022neurips-fast,
  title     = {{Fast Stochastic Composite Minimization and an Accelerated Frank-Wolfe Algorithm Under Parallelization}},
  author    = {Dubois-Taine, Benjamin and Bach, Francis R. and Berthet, Quentin and Taylor, Adrien},
  booktitle = {Neural Information Processing Systems},
  year      = {2022},
  url       = {https://mlanthology.org/neurips/2022/duboistaine2022neurips-fast/}
}