One-Step Diffusion Distillation via Deep Equilibrium Models
Abstract
Diffusion models excel at producing high-quality samples but naively require hundreds of iterations, prompting multiple attempts to distill this process into a faster network. Existing approaches, however, often require complex multi-stage distillation and perform sub-optimally in single-step image generation. In response, we introduce a simple yet effective means of diffusion distillation---*directly* mapping initial noise to the resulting image. Of particular importance to our approach is to leverage a new Deep Equilibrium (DEQ) model for distillation: the Generative Equilibrium Transformer (GET). Our method enables fully offline training with just noise/image pairs from the diffusion model while achieving superior performance compared to existing one-step methods on comparable training budgets. The DEQ architecture proves crucial, as GET matches a $5\times$ larger ViT in terms of FID scores while striking a critical balance of computational cost and image quality. Code, checkpoints, and datasets will be released.
Cite
Text
Geng et al. "One-Step Diffusion Distillation via Deep Equilibrium Models." ICML 2023 Workshops: DeployableGenerativeAI, 2023.Markdown
[Geng et al. "One-Step Diffusion Distillation via Deep Equilibrium Models." ICML 2023 Workshops: DeployableGenerativeAI, 2023.](https://mlanthology.org/icmlw/2023/geng2023icmlw-onestep/)BibTeX
@inproceedings{geng2023icmlw-onestep,
title = {{One-Step Diffusion Distillation via Deep Equilibrium Models}},
author = {Geng, Zhengyang and Pokle, Ashwini and Kolter, J Zico},
booktitle = {ICML 2023 Workshops: DeployableGenerativeAI},
year = {2023},
url = {https://mlanthology.org/icmlw/2023/geng2023icmlw-onestep/}
}