Asymptotic Analysis of Conditioned Stochastic Gradient Descent

Abstract

In this paper, we investigate a general class of stochastic gradient descent (SGD) algorithms, called $\textit{conditioned}$ SGD, based on a preconditioning of the gradient direction. Using a discrete-time approach with martingale tools, we establish under mild assumptions the weak convergence of the rescaled sequence of iterates for a broad class of conditioning matrices including stochastic first-order and second-order methods. Almost sure convergence results, which may be of independent interest, are also presented. Interestingly, the asymptotic normality result consists in a stochastic equicontinuity property so when the conditioning matrix is an estimate of the inverse Hessian, the algorithm is asymptotically optimal.

Cite

Text

Leluc and Portier. "Asymptotic Analysis of Conditioned Stochastic Gradient Descent." Transactions on Machine Learning Research, 2023.

Markdown

[Leluc and Portier. "Asymptotic Analysis of Conditioned Stochastic Gradient Descent." Transactions on Machine Learning Research, 2023.](https://mlanthology.org/tmlr/2023/leluc2023tmlr-asymptotic/)

BibTeX

@article{leluc2023tmlr-asymptotic,
  title     = {{Asymptotic Analysis of Conditioned Stochastic Gradient Descent}},
  author    = {Leluc, Rémi and Portier, François},
  journal   = {Transactions on Machine Learning Research},
  year      = {2023},
  url       = {https://mlanthology.org/tmlr/2023/leluc2023tmlr-asymptotic/}
}