3D Hand Pose Detection in Egocentric RGB-D Images

Abstract

We focus on the task of hand pose estimation from egocentric viewpoints. For this problem specification, we show that depth sensors are particularly informative for extracting near-field interactions of the camera wearer with his/her environment. Despite the recent advances in full-body pose estimation using Kinect-like sensors, reliable monocular hand pose estimation in RGB-D images is still an unsolved problem. The problem is exacerbated when considering a wearable sensor and a first-person camera viewpoint: the occlusions inherent to the particular camera view and the limitations in terms of field of view make the problem even more difficult. We propose to use task and viewpoint specific synthetic training exemplars in a discriminative detection framework. We also exploit the depth features for a sparser and faster detection. We evaluate our approach on a real-world annotated dataset and propose a novel annotation technique for accurate 3D hand labelling even in case of partial occlusions.

Cite

Text

Rogez et al. "3D Hand Pose Detection in Egocentric RGB-D Images." European Conference on Computer Vision Workshops, 2014. doi:10.1007/978-3-319-16178-5_25

Markdown

[Rogez et al. "3D Hand Pose Detection in Egocentric RGB-D Images." European Conference on Computer Vision Workshops, 2014.](https://mlanthology.org/eccvw/2014/rogez2014eccvw-3d/) doi:10.1007/978-3-319-16178-5_25

BibTeX

@inproceedings{rogez2014eccvw-3d,
  title     = {{3D Hand Pose Detection in Egocentric RGB-D Images}},
  author    = {Rogez, Grégory and Khademi, Maryam and Iii, James Steven Supancic and Montiel, J. M. M. and Ramanan, Deva},
  booktitle = {European Conference on Computer Vision Workshops},
  year      = {2014},
  pages     = {356-371},
  doi       = {10.1007/978-3-319-16178-5_25},
  url       = {https://mlanthology.org/eccvw/2014/rogez2014eccvw-3d/}
}