Expanded Parts Model for Human Attribute and Action Recognition in Still Images

Abstract

We propose a new model for recognizing human attributes (e.g. wearing a suit, sitting, short hair) and actions (e.g. running, riding a horse) in still images. The proposed model relies on a collection of part templates which are learnt discriminatively to explain specific scale-space locations in the images (in human centric coordinates). It avoids the limitations of highly structured models, which consist of a few (i.e. a mixture of) 'average' templates. To learn our model, we propose an algorithm which automatically mines out parts and learns corresponding discriminative templates with their respective locations from a large number of candidate parts. We validate the method on recent challenging datasets: (i) Willow 7 actions [7], (ii) 27 Human Attributes (HAT) [25], and (iii) Stanford 40 actions [37]. We obtain convincing qualitative and state-of-the-art quantitative results on the three datasets.

Cite

Text

Sharma et al. "Expanded Parts Model for Human Attribute and Action Recognition in Still Images." Conference on Computer Vision and Pattern Recognition, 2013. doi:10.1109/CVPR.2013.90

Markdown

[Sharma et al. "Expanded Parts Model for Human Attribute and Action Recognition in Still Images." Conference on Computer Vision and Pattern Recognition, 2013.](https://mlanthology.org/cvpr/2013/sharma2013cvpr-expanded/) doi:10.1109/CVPR.2013.90

BibTeX

@inproceedings{sharma2013cvpr-expanded,
  title     = {{Expanded Parts Model for Human Attribute and Action Recognition in Still Images}},
  author    = {Sharma, Gaurav and Jurie, Frederic and Schmid, Cordelia},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2013},
  doi       = {10.1109/CVPR.2013.90},
  url       = {https://mlanthology.org/cvpr/2013/sharma2013cvpr-expanded/}
}