DPHMs: Diffusion Parametric Head Models for Depth-Based Tracking
Abstract
We introduce Diffusion Parametric Head Models (DPHMs) a generative model that enables robust volumetric head reconstruction and tracking from monocular depth sequences. While recent volumetric head models such as NPHMs can now excel in representing high-fidelity head geometries tracking and reconstructing heads from real-world single-view depth sequences remains very challenging as the fitting to partial and noisy observations is underconstrained. To tackle these challenges we propose a latent diffusion-based prior to regularize volumetric head reconstruction and tracking. This prior-based regularizer effectively constrains the identity and expression codes to lie on the underlying latent manifold which represents plausible head shapes. To evaluate the effectiveness of the diffusion-based prior we collect a dataset of monocular Kinect sequences consisting of various complex facial expression motions and rapid transitions. We compare our method to state-of-the-art tracking methods and demonstrate improved head identity reconstruction as well as robust expression tracking.
Cite
Text
Tang et al. "DPHMs: Diffusion Parametric Head Models for Depth-Based Tracking." Conference on Computer Vision and Pattern Recognition, 2024. doi:10.1109/CVPR52733.2024.00112Markdown
[Tang et al. "DPHMs: Diffusion Parametric Head Models for Depth-Based Tracking." Conference on Computer Vision and Pattern Recognition, 2024.](https://mlanthology.org/cvpr/2024/tang2024cvpr-dphms/) doi:10.1109/CVPR52733.2024.00112BibTeX
@inproceedings{tang2024cvpr-dphms,
title = {{DPHMs: Diffusion Parametric Head Models for Depth-Based Tracking}},
author = {Tang, Jiapeng and Dai, Angela and Nie, Yinyu and Markhasin, Lev and Thies, Justus and Nießner, Matthias},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2024},
pages = {1111-1122},
doi = {10.1109/CVPR52733.2024.00112},
url = {https://mlanthology.org/cvpr/2024/tang2024cvpr-dphms/}
}