Learning Visual Symbols for Parsing Human Poses in Images
Abstract
Parsing human poses in images is fundamental in extracting critical visual information for artificial intelligent agents. Our goal is to learn self-contained body part representations from images, which we call visual symbols, and their symbol-wise geometric contexts in this parsing process. Each symbol is individually learned by categorizing visual features leveraged by geometric information. In the categorization, we use Latent Support Vector Machine followed by an efficient cross validation procedure. Then, these symbols naturally define geometric contexts of body parts in a fine granularity for effective inference. When the structure of the compositional parts is a tree, we derive an efficient approach to estimating human poses in images. Experiments on two large datasets suggest our approach outperforms state of the art methods.
Cite
Text
Wang and Li. "Learning Visual Symbols for Parsing Human Poses in Images." International Joint Conference on Artificial Intelligence, 2013.Markdown
[Wang and Li. "Learning Visual Symbols for Parsing Human Poses in Images." International Joint Conference on Artificial Intelligence, 2013.](https://mlanthology.org/ijcai/2013/wang2013ijcai-learning/)BibTeX
@inproceedings{wang2013ijcai-learning,
title = {{Learning Visual Symbols for Parsing Human Poses in Images}},
author = {Wang, Fang and Li, Yi},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2013},
pages = {2510-2517},
url = {https://mlanthology.org/ijcai/2013/wang2013ijcai-learning/}
}