Pose-Guided Human Parsing by an AND/OR Graph Using Pose-Context Features

Abstract

Parsing human into semantic parts is crucial to human-centric analysis. In this paper, we propose a human parsing pipeline that uses pose cues, e.g., estimates of human joint locations, to provide pose-guided segment proposals for semantic parts. These segment proposals are ranked using standard appearance cues, deep-learned semantic feature, and a novel pose feature called pose-context. Then these proposals are selected and assembled using an And-Or graph to output a parse of the person. The And-Or graph is able to deal with large human appearance variability due to pose, choice of clothing, etc. We evaluate our approach on the popular Penn-Fudan pedestrian parsing dataset, showing that it significantly outperforms the state of the art, and perform diagnostics to demonstrate the effectiveness of different stages of our pipeline.

Cite

Text

Xia et al. "Pose-Guided Human Parsing by an AND/OR Graph Using Pose-Context Features." AAAI Conference on Artificial Intelligence, 2016. doi:10.1609/AAAI.V30I1.10460

Markdown

[Xia et al. "Pose-Guided Human Parsing by an AND/OR Graph Using Pose-Context Features." AAAI Conference on Artificial Intelligence, 2016.](https://mlanthology.org/aaai/2016/xia2016aaai-pose/) doi:10.1609/AAAI.V30I1.10460

BibTeX

@inproceedings{xia2016aaai-pose,
  title     = {{Pose-Guided Human Parsing by an AND/OR Graph Using Pose-Context Features}},
  author    = {Xia, Fangting and Zhu, Jun and Wang, Peng and Yuille, Alan L.},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2016},
  pages     = {3632-3640},
  doi       = {10.1609/AAAI.V30I1.10460},
  url       = {https://mlanthology.org/aaai/2016/xia2016aaai-pose/}
}