Hessian-Free High-Resolution Nesterov Acceleration for Sampling
Abstract
Nesterov’s Accelerated Gradient (NAG) for optimization has better performance than its continuous time limit (noiseless kinetic Langevin) when a finite step-size is employed (Shi et al., 2021). This work explores the sampling counterpart of this phenonemon and proposes a diffusion process, whose discretizations can yield accelerated gradient-based MCMC methods. More precisely, we reformulate the optimizer of NAG for strongly convex functions (NAG-SC) as a Hessian-Free High-Resolution ODE, change its high-resolution coefficient to a hyperparameter, inject appropriate noise, and discretize the resulting diffusion process. The acceleration effect of the new hyperparameter is quantified and it is not an artificial one created by time-rescaling. Instead, acceleration beyond underdamped Langevin in $W_2$ distance is quantitatively established for log-strongly-concave-and-smooth targets, at both the continuous dynamics level and the discrete algorithm level. Empirical experiments in both log-strongly-concave and multi-modal cases also numerically demonstrate this acceleration.
Cite
Text
Li et al. "Hessian-Free High-Resolution Nesterov Acceleration for Sampling." International Conference on Machine Learning, 2022.Markdown
[Li et al. "Hessian-Free High-Resolution Nesterov Acceleration for Sampling." International Conference on Machine Learning, 2022.](https://mlanthology.org/icml/2022/li2022icml-hessianfree/)BibTeX
@inproceedings{li2022icml-hessianfree,
title = {{Hessian-Free High-Resolution Nesterov Acceleration for Sampling}},
author = {Li, Ruilin and Zha, Hongyuan and Tao, Molei},
booktitle = {International Conference on Machine Learning},
year = {2022},
pages = {13125-13162},
volume = {162},
url = {https://mlanthology.org/icml/2022/li2022icml-hessianfree/}
}