Structure-Guided Adversarial Training of Diffusion Models
Abstract
Diffusion models have demonstrated exceptional efficacy in various generative applications. While existing models focus on minimizing a weighted sum of denoising score matching losses for data distribution modeling their training primarily emphasizes instance-level optimization overlooking valuable structural information within each mini-batch indicative of pair-wise relationships among samples. To address this limitation we introduce Structure-guided Adversarial training of Diffusion Models (SADM). In this pioneering approach we compel the model to learn manifold structures between samples in each training batch. To ensure the model captures authentic manifold structures in the data distribution we advocate adversarial training of the diffusion generator against a novel structure discriminator in a minimax game distinguishing real manifold structures from the generated ones. SADM substantially outperforms existing methods in image generation and cross-domain fine-tuning tasks across 12 datasets establishing a new state-of-the-art FID of 1.58 and 2.11 on ImageNet for class-conditional image generation at resolutions of 256x256 and 512x512 respectively.
Cite
Text
Yang et al. "Structure-Guided Adversarial Training of Diffusion Models." Conference on Computer Vision and Pattern Recognition, 2024. doi:10.1109/CVPR52733.2024.00693Markdown
[Yang et al. "Structure-Guided Adversarial Training of Diffusion Models." Conference on Computer Vision and Pattern Recognition, 2024.](https://mlanthology.org/cvpr/2024/yang2024cvpr-structureguided/) doi:10.1109/CVPR52733.2024.00693BibTeX
@inproceedings{yang2024cvpr-structureguided,
title = {{Structure-Guided Adversarial Training of Diffusion Models}},
author = {Yang, Ling and Qian, Haotian and Zhang, Zhilong and Liu, Jingwei and Cui, Bin},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2024},
pages = {7256-7266},
doi = {10.1109/CVPR52733.2024.00693},
url = {https://mlanthology.org/cvpr/2024/yang2024cvpr-structureguided/}
}