Stable Consistency Tuning: Understanding and Improving Consistency Models

Abstract

Diffusion models achieve high-quality generation but suffer from slow sampling due to their iterative denoising process. Consistency models offer a faster alternative with competitive performance, trained via consistency distillation from pretrained diffusion models or directly from raw data. We introduce a novel framework interpreting consistency models through a Markov Decision Process (MDP), framing their training as value estimation via Temporal Difference (TD) Learning. This perspective reveals limitations in existing training strategies. Building on Easy Consistency Tuning (ECT), we propose Stable Consistency Tuning (SCT), which enhances variance reduction using the score identity. SCT significantly improves performance on CIFAR-10 and ImageNet-64.

Cite

Text

Wang et al. "Stable Consistency Tuning: Understanding and Improving Consistency Models." ICLR 2025 Workshops: DeLTa, 2025.

Markdown

[Wang et al. "Stable Consistency Tuning: Understanding and Improving Consistency Models." ICLR 2025 Workshops: DeLTa, 2025.](https://mlanthology.org/iclrw/2025/wang2025iclrw-stable/)

BibTeX

@inproceedings{wang2025iclrw-stable,
  title     = {{Stable Consistency Tuning: Understanding and Improving Consistency Models}},
  author    = {Wang, Fu-Yun and Geng, Zhengyang and Li, Hongsheng},
  booktitle = {ICLR 2025 Workshops: DeLTa},
  year      = {2025},
  url       = {https://mlanthology.org/iclrw/2025/wang2025iclrw-stable/}
}