Expressive Telepresence via Modular Codec Avatars
Abstract
VR telepresence consists of interacting with another human in a virtual space represented by an avatar. Today most avatars are cartoon-like, but soon the technology will allow video-realistic ones. This paper aims in this direction and presents Modular Codec Avatars (MCA), a method to generate hyper-realistic faces driven by the cameras in the VR headset. MCA extends traditional Codec Avatars (CA) by replacing the holistic models with a learned modular representation. It is important to note that traditional person-specific CAs are learned from few training samples, and typically lack robustness as well as limited expressiveness when transferring facial expressions. MCAs solve these issues by learning a modulated adaptive blending of different facial components as well as an exemplar-based latent alignment. We demonstrate that MCA achieves improved expressiveness and robustness w.r.t to CA in a variety of real-world datasets and practical scenarios. Finally, we showcase new applications in VR telepresence enabled by the proposed model.
Cite
Text
Chu et al. "Expressive Telepresence via Modular Codec Avatars." Proceedings of the European Conference on Computer Vision (ECCV), 2020. doi:10.1007/978-3-030-58610-2_20Markdown
[Chu et al. "Expressive Telepresence via Modular Codec Avatars." Proceedings of the European Conference on Computer Vision (ECCV), 2020.](https://mlanthology.org/eccv/2020/chu2020eccv-expressive/) doi:10.1007/978-3-030-58610-2_20BibTeX
@inproceedings{chu2020eccv-expressive,
title = {{Expressive Telepresence via Modular Codec Avatars}},
author = {Chu, Hang and Ma, Shugao and De la Torre, Fernando and Fidler, Sanja and Sheikh, Yaser},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2020},
doi = {10.1007/978-3-030-58610-2_20},
url = {https://mlanthology.org/eccv/2020/chu2020eccv-expressive/}
}