Poselet Key-Framing: A Model for Human Activity Recognition

Abstract

In this paper, we develop a new model for recognizing human actions. An action is modeled as a very sparse sequence of temporally local discriminative keyframes collections of partial key-poses of the actor(s), depicting key states in the action sequence. We cast the learning of keyframes in a max-margin discriminative framework, where we treat keyframes as latent variables. This allows us to (jointly) learn a set of most discriminative keyframes while also learning the local temporal context between them. Keyframes are encoded using a spatially-localizable poselet-like representation with HoG and BoW components learned from weak annotations; we rely on structured SVM formulation to align our components and mine for hard negatives to boost localization performance. This results in a model that supports spatio-temporal localization and is insensitive to dropped frames or partial observations. We show classification performance that is competitive with the state of the art on the benchmark UT-Interaction dataset and illustrate that our model outperforms prior methods in an on-line streaming setting.

Cite

Text

Raptis and Sigal. "Poselet Key-Framing: A Model for Human Activity Recognition." Conference on Computer Vision and Pattern Recognition, 2013. doi:10.1109/CVPR.2013.342

Markdown

[Raptis and Sigal. "Poselet Key-Framing: A Model for Human Activity Recognition." Conference on Computer Vision and Pattern Recognition, 2013.](https://mlanthology.org/cvpr/2013/raptis2013cvpr-poselet/) doi:10.1109/CVPR.2013.342

BibTeX

@inproceedings{raptis2013cvpr-poselet,
  title     = {{Poselet Key-Framing: A Model for Human Activity Recognition}},
  author    = {Raptis, Michalis and Sigal, Leonid},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2013},
  doi       = {10.1109/CVPR.2013.342},
  url       = {https://mlanthology.org/cvpr/2013/raptis2013cvpr-poselet/}
}