Rotation-Invariant Mixed Graphical Model Network for 2D Hand Pose Estimation
Abstract
In this paper, we propose a new architecture named Rotation-invariant Mixed Graphical Model Network (R-MGMN) to solve the problem of 2D hand pose estimation from a monocular RGB image. By integrating a rotation net, the R-MGMN is invariant to rotations of the hand in the image. It also has a pool of graphical models, from which a combination of graphical models could be selected, conditioning on the input image. Belief propagation is performed on each graphical model separately, generating a set of marginal distributions, which are taken as the confidence maps of hand keypoint positions. Final confidence maps are obtained by aggregating these confidence maps together. We evaluate the R-MGMN on two public hand pose datasets. Experiment results show our model outperforms the state-of-the-art algorithm which is widely used in 2D hand pose estimation by a noticeable margin.
Cite
Text
Kong et al. "Rotation-Invariant Mixed Graphical Model Network for 2D Hand Pose Estimation." Winter Conference on Applications of Computer Vision, 2020.Markdown
[Kong et al. "Rotation-Invariant Mixed Graphical Model Network for 2D Hand Pose Estimation." Winter Conference on Applications of Computer Vision, 2020.](https://mlanthology.org/wacv/2020/kong2020wacv-rotationinvariant/)BibTeX
@inproceedings{kong2020wacv-rotationinvariant,
title = {{Rotation-Invariant Mixed Graphical Model Network for 2D Hand Pose Estimation}},
author = {Kong, Deying and Ma, Haoyu and Chen, Yifei and Xie, Xiaohui},
booktitle = {Winter Conference on Applications of Computer Vision},
year = {2020},
url = {https://mlanthology.org/wacv/2020/kong2020wacv-rotationinvariant/}
}