Learning Hierarchical Models of Complex Daily Activities from Annotated Videos

Abstract

Effective recognition of complex long-term activities is becoming an increasingly important task in artificial intelligence. In this paper, we propose a novel approach for building models of complex long-term activities. First, we automatically learn the hierarchical structure of activities by learning about the 'parent-child' relation of activity components from a video using the variability in annotations acquired using multiple annotators. This variability allows for extracting the inherent hierarchical structure of the activity in a video. We consolidate hierarchical structures of the same activity from different videos into a unified stochastic grammar describing the overall activity. We then describe an inference mechanism to interpret new instances of activities. We use three datasets, which have been annotated by multiple annotators, of daily activity videos to demonstrate the effectiveness of our system.

Cite

Text

Tayyub et al. "Learning Hierarchical Models of Complex Daily Activities from Annotated Videos." IEEE/CVF Winter Conference on Applications of Computer Vision, 2018. doi:10.1109/WACV.2018.00182

Markdown

[Tayyub et al. "Learning Hierarchical Models of Complex Daily Activities from Annotated Videos." IEEE/CVF Winter Conference on Applications of Computer Vision, 2018.](https://mlanthology.org/wacv/2018/tayyub2018wacv-learning/) doi:10.1109/WACV.2018.00182

BibTeX

@inproceedings{tayyub2018wacv-learning,
  title     = {{Learning Hierarchical Models of Complex Daily Activities from Annotated Videos}},
  author    = {Tayyub, Jawad and Hawasly, Majd and Hogg, David C. and Cohn, Anthony G.},
  booktitle = {IEEE/CVF Winter Conference on Applications of Computer Vision},
  year      = {2018},
  pages     = {1633-1641},
  doi       = {10.1109/WACV.2018.00182},
  url       = {https://mlanthology.org/wacv/2018/tayyub2018wacv-learning/}
}