Im2Hands: Learning Attentive Implicit Representation of Interacting Two-Hand Shapes

Abstract

We present Implicit Two Hands (Im2Hands), the first neural implicit representation of two interacting hands. Unlike existing methods on two-hand reconstruction that rely on a parametric hand model and/or low-resolution meshes, Im2Hands can produce fine-grained geometry of two hands with high hand-to-hand and hand-to-image coherency. To handle the shape complexity and interaction context between two hands, Im2Hands models the occupancy volume of two hands -- conditioned on an RGB image and coarse 3D keypoints -- by two novel attention-based modules responsible for (1) initial occupancy estimation and (2) context-aware occupancy refinement, respectively. Im2Hands first learns per-hand neural articulated occupancy in the canonical space designed for each hand using query-image attention. It then refines the initial two-hand occupancy in the posed space to enhance the coherency between the two hand shapes using query-anchor attention. In addition, we introduce an optional keypoint refinement module to enable robust two-hand shape estimation from predicted hand keypoints in a single-image reconstruction scenario. We experimentally demonstrate the effectiveness of Im2Hands on two-hand reconstruction in comparison to related methods, where ours achieves state-of-the-art results. Our code is publicly available at https://github.com/jyunlee/Im2Hands.

Cite

Text

Lee et al. "Im2Hands: Learning Attentive Implicit Representation of Interacting Two-Hand Shapes." Conference on Computer Vision and Pattern Recognition, 2023. doi:10.1109/CVPR52729.2023.02028

Markdown

[Lee et al. "Im2Hands: Learning Attentive Implicit Representation of Interacting Two-Hand Shapes." Conference on Computer Vision and Pattern Recognition, 2023.](https://mlanthology.org/cvpr/2023/lee2023cvpr-im2hands/) doi:10.1109/CVPR52729.2023.02028

BibTeX

@inproceedings{lee2023cvpr-im2hands,
  title     = {{Im2Hands: Learning Attentive Implicit Representation of Interacting Two-Hand Shapes}},
  author    = {Lee, Jihyun and Sung, Minhyuk and Choi, Honggyu and Kim, Tae-Kyun},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2023},
  pages     = {21169-21178},
  doi       = {10.1109/CVPR52729.2023.02028},
  url       = {https://mlanthology.org/cvpr/2023/lee2023cvpr-im2hands/}
}