SyncTweedies: A General Generative Framework Based on Synchronized Diffusions

Abstract

We introduce a general diffusion synchronization framework for generating diverse visual content, including ambiguous images, panorama images, 3D mesh textures, and 3D Gaussian splats textures, using a pretrained image diffusion model. We first present an analysis of various scenarios for synchronizing multiple diffusion processes through a canonical space. Based on the analysis, we introduce a synchronized diffusion method, SyncTweedies, which averages the outputs of Tweedie’s formula while conducting denoising in multiple instance spaces. Compared to previous work that achieves synchronization through finetuning, SyncTweedies is a zero-shot method that does not require any finetuning, preserving the rich prior of diffusion models trained on Internet-scale image datasets without overfitting to specific domains. We verify that SyncTweedies offers the broadest applicability to diverse applications and superior performance compared to the previous state-of-the-art for each application. Our project page is at https://synctweedies.github.io.

Cite

Text

Kim et al. "SyncTweedies: A General Generative Framework Based on Synchronized Diffusions." Neural Information Processing Systems, 2024. doi:10.52202/079017-3016

Markdown

[Kim et al. "SyncTweedies: A General Generative Framework Based on Synchronized Diffusions." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/kim2024neurips-synctweedies/) doi:10.52202/079017-3016

BibTeX

@inproceedings{kim2024neurips-synctweedies,
  title     = {{SyncTweedies: A General Generative Framework Based on Synchronized Diffusions}},
  author    = {Kim, Jaihoon and Koo, Juil and Yeo, Kyeongmin and Sung, Minhyuk},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-3016},
  url       = {https://mlanthology.org/neurips/2024/kim2024neurips-synctweedies/}
}