Video-Based Human Action Classi.cation with Ambiguous Correspondences

Abstract

This paper describes a combined tracking-classification framework for the unsupervised classification of human action. While most existing approaches assume that featurewise correspondences on people are either available or not at all, this method explicitly formalizes how the probability of correspondences can be used in computation when the correspondences are ambiguous. It is also able to exploit in a probabilistic manner any foreground-background preprocessed segmentation, even if the segmentation is of low confidence. A principled analysis of the problem leads to a novel probabilistic action representation called the correspondence-ambiguous feature histogram array (CAFHA) that is robust to variations across similar actions. Our results show that the new framework outperforms the recent Zelnik-Manor and Irani method [19] for unsupervised event classi?cation. Additionally, the framework is extended to quasi real-time action inference, achieving good recognition accuracy despite changes in person identity and variations in the actions.

Cite

Text

Feng and Cham. "Video-Based Human Action Classi.cation with Ambiguous Correspondences." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2005. doi:10.1109/CVPR.2005.549

Markdown

[Feng and Cham. "Video-Based Human Action Classi.cation with Ambiguous Correspondences." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2005.](https://mlanthology.org/cvprw/2005/feng2005cvprw-videobased/) doi:10.1109/CVPR.2005.549

BibTeX

@inproceedings{feng2005cvprw-videobased,
  title     = {{Video-Based Human Action Classi.cation with Ambiguous Correspondences}},
  author    = {Feng, Zhou and Cham, Tat-Jen},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2005},
  pages     = {82},
  doi       = {10.1109/CVPR.2005.549},
  url       = {https://mlanthology.org/cvprw/2005/feng2005cvprw-videobased/}
}