Improved Training Technique for Shortcut Models

Abstract

Shortcut models represent a promising, non-adversarial paradigm for generative modeling, uniquely supporting one-step, few-step, and multi-step sampling from a single trained network. However, their widespread adoption has been stymied by critical performance bottlenecks. This paper tackles the five core issues that held shortcut models back: (1) the hidden flaw of compounding guidance, which we are the first to formalize, causing severe image artifacts; (2) inflexible fixed guidance that restricts inference-time control; (3) a pervasive frequency bias driven by a reliance on low-level distances in the direct domain, which biases reconstructions toward low frequencies; (4) divergent self-consistency arising from a conflict with EMA training; and (5) curvy flow trajectories that impede convergence. To address these challenges, we introduce iSM, a unified training framework that systematically resolves each limitation. Our framework is built on four key improvements: Intrinsic Guidance provides explicit, dynamic control over guidance strength, resolving both compounding guidance and inflexibility. A Multi-Level Wavelet Loss mitigates frequency bias to restore high-frequency details. Scaling Optimal Transport (sOT) reduces training variance and learns straighter, more stable generative paths. Finally, a Twin EMA strategy reconciles training stability with self-consistency. Extensive experiments on ImageNet 256x256 demonstrate that our approach yields substantial FID improvements over baseline shortcut models across one-step, few-step, and multi-step generation, making shortcut models a viable and competitive class of generative models.

Cite

Text

Nguyen et al. "Improved Training Technique for Shortcut Models." Advances in Neural Information Processing Systems, 2025.

Markdown

[Nguyen et al. "Improved Training Technique for Shortcut Models." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/nguyen2025neurips-improved/)

BibTeX

@inproceedings{nguyen2025neurips-improved,
  title     = {{Improved Training Technique for Shortcut Models}},
  author    = {Nguyen, Anh and Van Nguyen, Viet and Vu, Duc and Dao, Trung Tuan and Tran, Chi and Tran, Toan and Tran, Anh Tuan},
  booktitle = {Advances in Neural Information Processing Systems},
  year      = {2025},
  url       = {https://mlanthology.org/neurips/2025/nguyen2025neurips-improved/}
}