Generalizable One-Shot 3D Neural Head Avatar

Abstract

We present a method that reconstructs and animates a 3D head avatar from a single-view portrait image. Existing methods either involve time-consuming optimization for a specific person with multiple images, or they struggle to synthesize intricate appearance details beyond the facial region. To address these limitations, we propose a framework that not only generalizes to unseen identities based on a single-view image without requiring person-specific optimization, but also captures characteristic details within and beyond the face area (e.g. hairstyle, accessories, etc.). At the core of our method are three branches that produce three tri-planes representing the coarse 3D geometry, detailed appearance of a source image, as well as the expression of a target image. By applying volumetric rendering to the combination of the three tri-planes followed by a super-resolution module, our method yields a high fidelity image of the desired identity, expression and pose. Once trained, our model enables efficient 3D head avatar reconstruction and animation via a single forward pass through a network. Experiments show that the proposed approach generalizes well to unseen validation datasets, surpassing SOTA baseline methods by a large margin on head avatar reconstruction and animation.

Cite

Text

Li et al. "Generalizable One-Shot 3D Neural Head Avatar." Neural Information Processing Systems, 2023.

Markdown

[Li et al. "Generalizable One-Shot 3D Neural Head Avatar." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/li2023neurips-generalizable/)

BibTeX

@inproceedings{li2023neurips-generalizable,
  title     = {{Generalizable One-Shot 3D Neural Head Avatar}},
  author    = {Li, Xueting and De Mello, Shalini and Liu, Sifei and Nagano, Koki and Iqbal, Umar and Kautz, Jan},
  booktitle = {Neural Information Processing Systems},
  year      = {2023},
  url       = {https://mlanthology.org/neurips/2023/li2023neurips-generalizable/}
}