Path Planning for Masked Diffusion Models with Applications to Biological Sequence Generation
Abstract
In this paper, we investigate how the order in which tokens are unmasked during masked diffusion model (MDM) inference affects generative quality. We derive an expanded evidence lower bound (ELBO) that introduces a planner, responsible for selecting which tokens to unmask at each step. Our analysis suggests that alternative unmasking strategies can improve generative performance. Based on these insights, we propose Path Planning (P2), a training-free inference framework that leverages pre-trained BERT or the denoiser itself to guide unmasking decisions. P2 generalizes all known MDM sampling strategies and enables significant improvements across diverse domains including language generation (in-context learning, code generation, story infilling, mathematical reasoning, reverse curse correction) and biological sequence generation (protein and RNA sequences).
Cite
Text
Peng et al. "Path Planning for Masked Diffusion Models with Applications to Biological Sequence Generation." ICLR 2025 Workshops: DeLTa, 2025.Markdown
[Peng et al. "Path Planning for Masked Diffusion Models with Applications to Biological Sequence Generation." ICLR 2025 Workshops: DeLTa, 2025.](https://mlanthology.org/iclrw/2025/peng2025iclrw-path/)BibTeX
@inproceedings{peng2025iclrw-path,
title = {{Path Planning for Masked Diffusion Models with Applications to Biological Sequence Generation}},
author = {Peng, Fred Zhangzhi and Bezemek, Zachary and Patel, Sawan and Rector-Brooks, Jarrid and Yao, Sherwood and Tong, Alexander and Chatterjee, Pranam},
booktitle = {ICLR 2025 Workshops: DeLTa},
year = {2025},
url = {https://mlanthology.org/iclrw/2025/peng2025iclrw-path/}
}