Dominant Codewords Selection with Topic Model for Action Recognition

Abstract

In this paper, we propose a framework for recognizing human activities that uses only in-topic dominant codewords and a mixture of intertopic vectors. Latent Dirichlet allocation (LDA) is used to develop approximations of human motion primitives, these are mid-level representations, and they adaptively integrate dominant vectors when classifying human activities. In LDA topic modeling, action videos (documents) are represented by a bag-of-words (input from a dictionary), and these are based on improved dense trajectories ([18]). The output topics correspond to human motion primitives, such as finger moving or subtle leg motion. We eliminate the impurities, such as missed tracking or changing light conditions, in each motion primitive. The assembled vector of motion primitives is an improved representation of the action. We demonstrate our method on four different datasets.

Cite

Text

Kataoka et al. "Dominant Codewords Selection with Topic Model for Action Recognition." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2016. doi:10.1109/CVPRW.2016.101

Markdown

[Kataoka et al. "Dominant Codewords Selection with Topic Model for Action Recognition." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2016.](https://mlanthology.org/cvprw/2016/kataoka2016cvprw-dominant/) doi:10.1109/CVPRW.2016.101

BibTeX

@inproceedings{kataoka2016cvprw-dominant,
  title     = {{Dominant Codewords Selection with Topic Model for Action Recognition}},
  author    = {Kataoka, Hirokatsu and Iwata, Kenji and Satoh, Yutaka and Hayashi, Masaki and Aoki, Yoshimitsu and Ilic, Slobodan},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2016},
  pages     = {770-777},
  doi       = {10.1109/CVPRW.2016.101},
  url       = {https://mlanthology.org/cvprw/2016/kataoka2016cvprw-dominant/}
}