Accelerated Parameter-Free Stochastic Optimization

Abstract

We propose a method that achieves near-optimal rates for \emph{smooth} stochastic convex optimization and requires essentially no prior knowledge of problem parameters. This improves on prior work which requires knowing at least the initial distance to optimality $d_0$. Our method, \textsc{U-DoG}, combines \textsc{UniXGrad} (Kavis et al., 2019) and \textsc{DoG} (Ivgi et al., 2023) with novel iterate stabilization techniques. It requires only loose bounds on $d_0$ and the noise magnitude, provides high probability guarantees under sub-Gaussian noise, and is also near-optimal in the non-smooth case. Our experiments show consistent, strong performance on convex problems and mixed results on neural network training.

Cite

Text

Kreisler et al. "Accelerated Parameter-Free Stochastic Optimization." Conference on Learning Theory, 2024.

Markdown

[Kreisler et al. "Accelerated Parameter-Free Stochastic Optimization." Conference on Learning Theory, 2024.](https://mlanthology.org/colt/2024/kreisler2024colt-accelerated/)

BibTeX

@inproceedings{kreisler2024colt-accelerated,
  title     = {{Accelerated Parameter-Free Stochastic Optimization}},
  author    = {Kreisler, Itai and Ivgi, Maor and Hinder, Oliver and Carmon, Yair},
  booktitle = {Conference on Learning Theory},
  year      = {2024},
  pages     = {3257-3324},
  volume    = {247},
  url       = {https://mlanthology.org/colt/2024/kreisler2024colt-accelerated/}
}