Pluralistic Aging Diffusion Autoencoder

Abstract

Face aging is an ill-posed problem because multiple plausible aging patterns may correspond to a given input. Most existing methods often produce one deterministic estimation. This paper proposes a novel CLIP-driven Pluralistic Aging Diffusion Autoencoder (PADA) to enhance the diversity of aging patterns. First, we employ diffusion models to generate diverse low-level aging details via a sequential denoising reverse process. Second, we present Probabilistic Aging Embedding (PAE) to capture diverse high-level aging patterns, which represents age information as probabilistic distributions in the common CLIP latent space. A text-guided KL-divergence loss is designed to guide this learning. Our method can achieve pluralistic face aging conditioned on open-world aging texts and arbitrary unseen face images. Qualitative and quantitative experiments demonstrate that our method can generate more diverse and high-quality plausible aging results.

Cite

Text

Li et al. "Pluralistic Aging Diffusion Autoencoder." International Conference on Computer Vision, 2023. doi:10.1109/ICCV51070.2023.02067

Markdown

[Li et al. "Pluralistic Aging Diffusion Autoencoder." International Conference on Computer Vision, 2023.](https://mlanthology.org/iccv/2023/li2023iccv-pluralistic/) doi:10.1109/ICCV51070.2023.02067

BibTeX

@inproceedings{li2023iccv-pluralistic,
  title     = {{Pluralistic Aging Diffusion Autoencoder}},
  author    = {Li, Peipei and Wang, Rui and Huang, Huaibo and He, Ran and He, Zhaofeng},
  booktitle = {International Conference on Computer Vision},
  year      = {2023},
  pages     = {22613-22623},
  doi       = {10.1109/ICCV51070.2023.02067},
  url       = {https://mlanthology.org/iccv/2023/li2023iccv-pluralistic/}
}