Scalable and Efficient Probabilistic Inference for Bayesian Deep Learning and Generative Modeling
Abstract
Probabilistic inference is a fundamental challenge in machine learning, spanning tasks from approximate Bayesian inference to generative AI. In this talk, I will present theoretically-guaranteed scalable and efficient probabilistic inference with applications in Bayesian deep learning and generative modeling. First, I will introduce a new compute paradigm for probabilistic inference that leverages modern accelerators, specifically low-precision and sparsity, to significantly speed up inference while preserving accuracy. Next, I will present a new framework for efficient inference in discrete domains, utilizing gradient information—a largely overlooked feature of discrete distributions—to enable more informed and directional exploration. Finally, I will showcase experimental results demonstrating the effectiveness of these methods across various ML tasks, including Bayesian neural networks, energy-based models, and large language models.
Cite
Text
Zhang. "Scalable and Efficient Probabilistic Inference for Bayesian Deep Learning and Generative Modeling." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I27.35129Markdown
[Zhang. "Scalable and Efficient Probabilistic Inference for Bayesian Deep Learning and Generative Modeling." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/zhang2025aaai-scalable/) doi:10.1609/AAAI.V39I27.35129BibTeX
@inproceedings{zhang2025aaai-scalable,
title = {{Scalable and Efficient Probabilistic Inference for Bayesian Deep Learning and Generative Modeling}},
author = {Zhang, Ruqi},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2025},
pages = {28737},
doi = {10.1609/AAAI.V39I27.35129},
url = {https://mlanthology.org/aaai/2025/zhang2025aaai-scalable/}
}