Human Daily Action Analysis with Multi-View and Color-Depth Data

Abstract

Improving human action recognition in videos is restricted by the inherent limitations of the visual data. In this paper, we take the depth information into consideration and construct a novel dataset of human daily actions. The proposed ACT4^2 dataset provides synchronized data from 4 views and 2 sources, aiming to facilitate the research of action analysis across multiple views and multiple sources. We also propose a new descriptor of depth information for action representation, which depicts the structural relations of spatiotemporal points within action volume using the distance information in depth data. In experimental validation, our descriptor obtains superior performance to the state-of-the-art action descriptors designed for color information, and more robust to viewpoint variations. The fusion of features from different sources is also discussed, and a simple but efficient method is presented to provide a baseline performance on the proposed dataset.

Cite

Text

Cheng et al. "Human Daily Action Analysis with Multi-View and Color-Depth Data." European Conference on Computer Vision Workshops, 2012. doi:10.1007/978-3-642-33868-7_6

Markdown

[Cheng et al. "Human Daily Action Analysis with Multi-View and Color-Depth Data." European Conference on Computer Vision Workshops, 2012.](https://mlanthology.org/eccvw/2012/cheng2012eccvw-human/) doi:10.1007/978-3-642-33868-7_6

BibTeX

@inproceedings{cheng2012eccvw-human,
  title     = {{Human Daily Action Analysis with Multi-View and Color-Depth Data}},
  author    = {Cheng, Zhongwei and Qin, Lei and Ye, Yituo and Huang, Qingming and Tian, Qi},
  booktitle = {European Conference on Computer Vision Workshops},
  year      = {2012},
  pages     = {52-61},
  doi       = {10.1007/978-3-642-33868-7_6},
  url       = {https://mlanthology.org/eccvw/2012/cheng2012eccvw-human/}
}