Diffusion Autoencoders: Toward a Meaningful and Decodable Representation
Abstract
Diffusion probabilistic models (DPMs) have achieved remarkable quality in image generation that rivals GANs'. But unlike GANs, DPMs use a set of latent variables that lack semantic meaning and cannot serve as a useful representation for other tasks. This paper explores the possibility of using DPMs for representation learning and seeks to extract a meaningful and decodable representation of an input image via autoencoding. Our key idea is to use a learnable encoder for discovering the high-level semantics, and a DPM as the decoder for modeling the remaining stochastic variations. Our method can encode any image into a two-part latent code, where the first part is semantically meaningful and linear, and the second part captures stochastic details, allowing near-exact reconstruction. This capability enables challenging applications that currently foil GAN-based methods, such as attribute manipulation on real images. We also show that this two-level encoding improves denoising efficiency and naturally facilitates various downstream tasks including few-shot conditional sampling.
Cite
Text
Preechakul et al. "Diffusion Autoencoders: Toward a Meaningful and Decodable Representation." Conference on Computer Vision and Pattern Recognition, 2022. doi:10.1109/CVPR52688.2022.01036Markdown
[Preechakul et al. "Diffusion Autoencoders: Toward a Meaningful and Decodable Representation." Conference on Computer Vision and Pattern Recognition, 2022.](https://mlanthology.org/cvpr/2022/preechakul2022cvpr-diffusion/) doi:10.1109/CVPR52688.2022.01036BibTeX
@inproceedings{preechakul2022cvpr-diffusion,
title = {{Diffusion Autoencoders: Toward a Meaningful and Decodable Representation}},
author = {Preechakul, Konpat and Chatthee, Nattanat and Wizadwongsa, Suttisak and Suwajanakorn, Supasorn},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2022},
pages = {10619-10629},
doi = {10.1109/CVPR52688.2022.01036},
url = {https://mlanthology.org/cvpr/2022/preechakul2022cvpr-diffusion/}
}