Non Asymptotic Analysis of Adaptive Stochastic Gradient Algorithms and Applications

Abstract

In stochastic optimization, a widely used approach for handling large samples sequentially is the stochastic gradient algorithm (SGD). However, a key limitation of SGD is that its step size sequence remains uniform across all gradient directions, which can lead to poor performance in practice, particularly for ill-conditioned problems. To address this issue, adaptive gradient algorithms, such as Adagrad and stochastic Newton methods, have been developed. These algorithms adapt the step size to each gradient direction, providing significant advantages in such challenging settings. This paper focuses on the non-asymptotic analysis of these adaptive gradient algorithms for strongly convex objective functions. The theoretical results are further applied to practical examples, including linear regression and regularized generalized linear models, using both Adagrad and stochastic Newton algorithms.

Cite

Text

Godichon-Baggioni and Tarrago. "Non Asymptotic Analysis of Adaptive Stochastic Gradient Algorithms and Applications." Transactions on Machine Learning Research, 2025.

Markdown

[Godichon-Baggioni and Tarrago. "Non Asymptotic Analysis of Adaptive Stochastic Gradient Algorithms and Applications." Transactions on Machine Learning Research, 2025.](https://mlanthology.org/tmlr/2025/godichonbaggioni2025tmlr-non/)

BibTeX

@article{godichonbaggioni2025tmlr-non,
  title     = {{Non Asymptotic Analysis of Adaptive Stochastic Gradient Algorithms and Applications}},
  author    = {Godichon-Baggioni, Antoine and Tarrago, Pierre},
  journal   = {Transactions on Machine Learning Research},
  year      = {2025},
  url       = {https://mlanthology.org/tmlr/2025/godichonbaggioni2025tmlr-non/}
}