AE-StyleGAN: Improved Training of Style-Based Auto-Encoders

Abstract

StyleGANs have shown impressive results on data generation and manipulation in recent years, thanks to its disentangled style latent space. A lot of efforts have been made in inverting a pre-trained generator, where an encoder is trained ad hoc after the generator is trained in a two-stage fashion. In this paper, we focus on style-based generators asking a scientific question: Does forcing such a generator to reconstruct real data lead to more disentangled latent space and make the inversion process from image to latent space easy? We describe a new methodology to train a style-based autoencoder where the encoder and generator are optimized end-to-end. We show that our proposed model consistently outperforms baselines in terms of image inversion and generation quality.

Cite

Text

Han et al. "AE-StyleGAN: Improved Training of Style-Based Auto-Encoders." Winter Conference on Applications of Computer Vision, 2022.

Markdown

[Han et al. "AE-StyleGAN: Improved Training of Style-Based Auto-Encoders." Winter Conference on Applications of Computer Vision, 2022.](https://mlanthology.org/wacv/2022/han2022wacv-aestylegan/)

BibTeX

@inproceedings{han2022wacv-aestylegan,
  title     = {{AE-StyleGAN: Improved Training of Style-Based Auto-Encoders}},
  author    = {Han, Ligong and Musunuri, Sri Harsha and Min, Martin Renqiang and Gao, Ruijiang and Tian, Yu and Metaxas, Dimitris},
  booktitle = {Winter Conference on Applications of Computer Vision},
  year      = {2022},
  pages     = {3134-3143},
  url       = {https://mlanthology.org/wacv/2022/han2022wacv-aestylegan/}
}