Gradient-Based Discrete Sampling with Automatic Cyclical Scheduling
Abstract
Discrete distributions are often highly multimodal due to discontinuities inherent in their spaces, especially in high-dimensional distributions encountered in deep models. While gradient-based discrete sampling has proven effective for sampling in discrete spaces, it is susceptible to becoming trapped in local modes due to the gradient information. To tackle this challenge, we propose an automatic cyclical scheduling, designed for efficient and accurate sampling in multimodal discrete distributions. Our method contains three key components: (1) a cyclical step size schedule where large steps discover new modes and small steps exploit each mode; (2) a cyclical balancing schedule, ensuring "balanced" proposals for a given step size; and (3) an automatic tuning scheme for adjusting the hyperparameters in the cyclical schedules, allowing adaptability across diverse datasets without manual tuning. We prove the non-asymptotic convergence of our method in general discrete distributions. Various experiments demonstrate the superiority of our method in learning complex multimodal discrete distributions.
Cite
Text
Pynadath et al. "Gradient-Based Discrete Sampling with Automatic Cyclical Scheduling." ICML 2024 Workshops: SPIGM, 2024.Markdown
[Pynadath et al. "Gradient-Based Discrete Sampling with Automatic Cyclical Scheduling." ICML 2024 Workshops: SPIGM, 2024.](https://mlanthology.org/icmlw/2024/pynadath2024icmlw-gradientbased/)BibTeX
@inproceedings{pynadath2024icmlw-gradientbased,
title = {{Gradient-Based Discrete Sampling with Automatic Cyclical Scheduling}},
author = {Pynadath, Patrick and Bhattacharya, Riddhiman and Hariharan, Arun Narayanan and Zhang, Ruqi},
booktitle = {ICML 2024 Workshops: SPIGM},
year = {2024},
url = {https://mlanthology.org/icmlw/2024/pynadath2024icmlw-gradientbased/}
}