Modeling Supporting Regions for Close Human Interaction Recognition

Abstract

This paper addresses the problem of recognizing human interactions with close physical contact from videos. Different from conventional human interaction recognition, recognizing close interactions faces the problems of ambiguities in feature-to-person assignments and frequent occlusions. Therefore, it is infeasible to accurately extract the interacting people, and the recognition performance of an interaction model is degraded. We propose a patch-aware model to overcome the two problems in close interaction recognition. Our model learns discriminative supporting regions for each interacting individual. The learned supporting regions accurately extract individuals at patch level, and explicitly indicate feature assignments. In addition, our model encodes a set of body part configurations for one interaction class, which provide rich representations for frequent occlusions. Our approach is evaluated on the UT-Interaction dataset and the BIT-Interaction dataset, and achieves promising results.

Cite

Text

Kong and Fu. "Modeling Supporting Regions for Close Human Interaction Recognition." European Conference on Computer Vision Workshops, 2014. doi:10.1007/978-3-319-16181-5_3

Markdown

[Kong and Fu. "Modeling Supporting Regions for Close Human Interaction Recognition." European Conference on Computer Vision Workshops, 2014.](https://mlanthology.org/eccvw/2014/kong2014eccvw-modeling/) doi:10.1007/978-3-319-16181-5_3

BibTeX

@inproceedings{kong2014eccvw-modeling,
  title     = {{Modeling Supporting Regions for Close Human Interaction Recognition}},
  author    = {Kong, Yu and Fu, Yun},
  booktitle = {European Conference on Computer Vision Workshops},
  year      = {2014},
  pages     = {29-44},
  doi       = {10.1007/978-3-319-16181-5_3},
  url       = {https://mlanthology.org/eccvw/2014/kong2014eccvw-modeling/}
}