Optimal Dimension Dependence of the Metropolis-Adjusted Langevin Algorithm

Abstract

Conventional wisdom in the sampling literature, backed by a popular diffusion scaling limit, suggests that the mixing time of the Metropolis-Adjusted Langevin Algorithm (MALA) scales as O(d^1/3), where d is the dimension. However, the diffusion scaling limit requires stringent assumptions on the target distribution and is asymptotic in nature. In contrast, the best known non-asymptotic mixing time bound for MALA on the class of log-smooth and strongly log-concave distributions is O(d). In this work, we establish that the mixing time of MALA on this class of target distributions is \tilde\Theta(d^1/2) under a warm start. Our upper bound proof introduces a new technique based on a projection characterization of the Metropolis adjustment which reduces the study of MALA to the well-studied discretization analysis of the Langevin SDE and bypasses direct computation of the acceptance probability.

Cite

Text

Chewi et al. "Optimal Dimension Dependence of the Metropolis-Adjusted Langevin Algorithm." Conference on Learning Theory, 2021.

Markdown

[Chewi et al. "Optimal Dimension Dependence of the Metropolis-Adjusted Langevin Algorithm." Conference on Learning Theory, 2021.](https://mlanthology.org/colt/2021/chewi2021colt-optimal/)

BibTeX

@inproceedings{chewi2021colt-optimal,
  title     = {{Optimal Dimension Dependence of the Metropolis-Adjusted Langevin Algorithm}},
  author    = {Chewi, Sinho and Lu, Chen and Ahn, Kwangjun and Cheng, Xiang and Le Gouic, Thibaut and Rigollet, Philippe},
  booktitle = {Conference on Learning Theory},
  year      = {2021},
  pages     = {1260-1300},
  volume    = {134},
  url       = {https://mlanthology.org/colt/2021/chewi2021colt-optimal/}
}