Sparse-View 3D Reconstruction of Clothed Humans via Normal Maps

Abstract

We present a novel deep learning-based approach to the 3D reconstruction of clothed humans using weak supervision via 2D normal maps. Given a single RGB image or multiview images our network is optimized to infer a person-specific signed distance function (SDF) discretized on a tetrahedral mesh surrounding the body in a rest pose. Subsequently estimated human pose and camera parameters are used to generate a normal map from the SDF. A key aspect of our approach is the direct use of the Marching Tetrahedra algorithm in end-to-end optimization and in order to do so we derive analytical gradients to facilitate straightforward differentiation (and thus backpropagation). Additionally predicted normal maps allow us to leverage pretrained image-to-normal networks in order to minimize a surface error instead of a photometric error. We demonstrate the efficacy of our approach on both labeled and in-the-wild data in the context of existing clothed human reconstruction methods.

Cite

Text

Wu et al. "Sparse-View 3D Reconstruction of Clothed Humans via Normal Maps." Winter Conference on Applications of Computer Vision, 2025.

Markdown

[Wu et al. "Sparse-View 3D Reconstruction of Clothed Humans via Normal Maps." Winter Conference on Applications of Computer Vision, 2025.](https://mlanthology.org/wacv/2025/wu2025wacv-sparseview/)

BibTeX

@inproceedings{wu2025wacv-sparseview,
  title     = {{Sparse-View 3D Reconstruction of Clothed Humans via Normal Maps}},
  author    = {Wu, Jane and Thomas, Diego and Fedkiw, Ronald},
  booktitle = {Winter Conference on Applications of Computer Vision},
  year      = {2025},
  pages     = {11-22},
  url       = {https://mlanthology.org/wacv/2025/wu2025wacv-sparseview/}
}