Revisiting Diffusion Models: From Generative Pre-Training to One-Step Generation

Abstract

Diffusion distillation is a widely used technique to reduce the sampling cost of diffusion models, yet it often requires extensive training, and the student performance tends to be degraded. Recent studies show that incorporating a GAN objective may alleviate these issues, yet the underlying mechanism remains unclear. In this work, we first identify a key limitation of distillation: mismatched step sizes and parameter numbers between the teacher and the student model lead them to converge to different local minima, rendering direct imitation suboptimal. We further demonstrate that a standalone GAN objective, without relying a distillation loss, overcomes this limitation and is sufficient to convert diffusion models into efficient one-step generators. Based on this finding, we propose that diffusion training may be viewed as a form of generative pre-training, equipping models with capabilities that can be unlocked through lightweight GAN fine-tuning. Supporting this view, we create a one-step generation model by fine-tuning a pre-trained model with 85% of parameters frozen, achieving strong performance with only 0.2M images and near-SOTA results with 5M images. We further present a frequency-domain analysis that may explain the one-step generative capability gained in diffusion training. Overall, our work provides a new perspective for diffusion training, highlighting its role as a powerful generative pre-training process, which can be the basis for building efficient one-step generation models.

Cite

Text

Zheng and Yang. "Revisiting Diffusion Models: From Generative Pre-Training to One-Step Generation." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[Zheng and Yang. "Revisiting Diffusion Models: From Generative Pre-Training to One-Step Generation." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/zheng2025icml-revisiting/)

BibTeX

@inproceedings{zheng2025icml-revisiting,
  title     = {{Revisiting Diffusion Models: From Generative Pre-Training to One-Step Generation}},
  author    = {Zheng, Bowen and Yang, Tianming},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {78434-78453},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/zheng2025icml-revisiting/}
}