SPMDM: Enhancing Masked Diffusion Models Through Simplifying Sampling Path
Abstract
Autoregressive models (ARMs) show strong capabilities in many domains but face challenges with planning and complex reasoning due to their sequential generation. Masked diffusion models (MDMs) address these issues by enabling controllable, any-order, and parallel generation but encounter training difficulties as token prediction complexity varies with unmasked token positions. This work identifies two key characteristics of efficient MDM sampling paths: prioritizing tokens near unmasked ones and generating subsequence earlier in reasoning. We propose the Simple Path Masked Diffusion Model (SPMDM), which partitions sequences into fixed-length, non-overlapping subsequences and applies varying noise scales to learn token-level and cross-subsequence dependencies. Experiments on synthetic data and tasks like Countdown and Sudoku show SPMDM captures structural rules effectively, significantly outperforming existing MDMs and ARMs, with competitive results on broader reasoning benchmarks.
Cite
Text
Zhu et al. "SPMDM: Enhancing Masked Diffusion Models Through Simplifying Sampling Path." Advances in Neural Information Processing Systems, 2025.Markdown
[Zhu et al. "SPMDM: Enhancing Masked Diffusion Models Through Simplifying Sampling Path." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/zhu2025neurips-spmdm/)BibTeX
@inproceedings{zhu2025neurips-spmdm,
title = {{SPMDM: Enhancing Masked Diffusion Models Through Simplifying Sampling Path}},
author = {Zhu, Yichen and Chen, Weiyu and Kwok, James and Zhao, Zhou},
booktitle = {Advances in Neural Information Processing Systems},
year = {2025},
url = {https://mlanthology.org/neurips/2025/zhu2025neurips-spmdm/}
}