Flash Diffusion: Accelerating Any Conditional Diffusion Model for Few Steps Image Generation

Abstract

In this paper, we propose an efficient, fast, and versatile distillation method to accelerate the generation of pre-trained diffusion models. The method reaches state-of-the-art performances in terms of FID and CLIP-Score for few steps image generation on the COCO2014 and COCO2017 datasets, while requiring only several GPU hours of training and fewer trainable parameters than existing methods. In addition to its efficiency, the versatility of the method is also exposed across several tasks such as *text-to-image*, *inpainting*, *face-swapping*, *super-resolution* and using different backbones such as UNet-based denoisers (SD1.5, SDXL), DiT (Pixart) and MMDiT (SD3), as well as adapters. In all cases, the method allowed to reduce drastically the number of sampling steps while maintaining very high-quality image generation.

Cite

Text

Chadebec et al. "Flash Diffusion: Accelerating Any Conditional Diffusion Model for Few Steps Image Generation." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I15.33722

Markdown

[Chadebec et al. "Flash Diffusion: Accelerating Any Conditional Diffusion Model for Few Steps Image Generation." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/chadebec2025aaai-flash/) doi:10.1609/AAAI.V39I15.33722

BibTeX

@inproceedings{chadebec2025aaai-flash,
  title     = {{Flash Diffusion: Accelerating Any Conditional Diffusion Model for Few Steps Image Generation}},
  author    = {Chadebec, Clément and Tasar, Onur and Benaroche, Eyal and Aubin, Benjamin},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {15686-15695},
  doi       = {10.1609/AAAI.V39I15.33722},
  url       = {https://mlanthology.org/aaai/2025/chadebec2025aaai-flash/}
}