Hessian-Free High-Resolution Nesterov Acceleration for Sampling

Abstract

Nesterov’s Accelerated Gradient (NAG) for optimization has better performance than its continuous time limit (noiseless kinetic Langevin) when a finite step-size is employed (Shi et al., 2021). This work explores the sampling counterpart of this phenonemon and proposes a diffusion process, whose discretizations can yield accelerated gradient-based MCMC methods. More precisely, we reformulate the optimizer of NAG for strongly convex functions (NAG-SC) as a Hessian-Free High-Resolution ODE, change its high-resolution coefficient to a hyperparameter, inject appropriate noise, and discretize the resulting diffusion process. The acceleration effect of the new hyperparameter is quantified and it is not an artificial one created by time-rescaling. Instead, acceleration beyond underdamped Langevin in $W_2$ distance is quantitatively established for log-strongly-concave-and-smooth targets, at both the continuous dynamics level and the discrete algorithm level. Empirical experiments in both log-strongly-concave and multi-modal cases also numerically demonstrate this acceleration.

Cite

Text

Li et al. "Hessian-Free High-Resolution Nesterov Acceleration for Sampling." International Conference on Machine Learning, 2022.

Markdown

[Li et al. "Hessian-Free High-Resolution Nesterov Acceleration for Sampling." International Conference on Machine Learning, 2022.](https://mlanthology.org/icml/2022/li2022icml-hessianfree/)

BibTeX

@inproceedings{li2022icml-hessianfree,
  title     = {{Hessian-Free High-Resolution Nesterov Acceleration for Sampling}},
  author    = {Li, Ruilin and Zha, Hongyuan and Tao, Molei},
  booktitle = {International Conference on Machine Learning},
  year      = {2022},
  pages     = {13125-13162},
  volume    = {162},
  url       = {https://mlanthology.org/icml/2022/li2022icml-hessianfree/}
}