3D Hand Pose Detection in Egocentric RGB-D Images
Abstract
We focus on the task of hand pose estimation from egocentric viewpoints. For this problem specification, we show that depth sensors are particularly informative for extracting near-field interactions of the camera wearer with his/her environment. Despite the recent advances in full-body pose estimation using Kinect-like sensors, reliable monocular hand pose estimation in RGB-D images is still an unsolved problem. The problem is exacerbated when considering a wearable sensor and a first-person camera viewpoint: the occlusions inherent to the particular camera view and the limitations in terms of field of view make the problem even more difficult. We propose to use task and viewpoint specific synthetic training exemplars in a discriminative detection framework. We also exploit the depth features for a sparser and faster detection. We evaluate our approach on a real-world annotated dataset and propose a novel annotation technique for accurate 3D hand labelling even in case of partial occlusions.
Cite
Text
Rogez et al. "3D Hand Pose Detection in Egocentric RGB-D Images." European Conference on Computer Vision Workshops, 2014. doi:10.1007/978-3-319-16178-5_25Markdown
[Rogez et al. "3D Hand Pose Detection in Egocentric RGB-D Images." European Conference on Computer Vision Workshops, 2014.](https://mlanthology.org/eccvw/2014/rogez2014eccvw-3d/) doi:10.1007/978-3-319-16178-5_25BibTeX
@inproceedings{rogez2014eccvw-3d,
title = {{3D Hand Pose Detection in Egocentric RGB-D Images}},
author = {Rogez, Grégory and Khademi, Maryam and Iii, James Steven Supancic and Montiel, J. M. M. and Ramanan, Deva},
booktitle = {European Conference on Computer Vision Workshops},
year = {2014},
pages = {356-371},
doi = {10.1007/978-3-319-16178-5_25},
url = {https://mlanthology.org/eccvw/2014/rogez2014eccvw-3d/}
}