Functional Categorization of Objects Using Real-Time Markerless Motion Capture

Abstract

Unsupervised categorization of objects is a fundamental problem in computer vision. While appearance-based methods have become popular recently, other important cues like functionality are largely neglected. Motivated by psychological studies giving evidence that human demonstration has a facilitative effect on categorization in infancy, we propose an approach for object categorization from depth video streams. To this end, we have developed a method for capturing human motion in real-time. The captured data is then used to temporally segment the depth streams into actions. The set of segmented actions are then categorized in an un-supervised manner, through a novel descriptor for motion capture data that is robust to subject variations. Furthermore, we automatically localize the object that is manipulated within a video segment, and categorize it using the corresponding action. For evaluation, we have recorded a dataset that comprises depth data with registered video sequences for 6 subjects, 13 action classes, and 174 object manipulations. © 2011 IEEE.

Cite

Text

Gall et al. "Functional Categorization of Objects Using Real-Time Markerless Motion Capture." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2011. doi:10.1109/CVPR.2011.5995582

Markdown

[Gall et al. "Functional Categorization of Objects Using Real-Time Markerless Motion Capture." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2011.](https://mlanthology.org/cvpr/2011/gall2011cvpr-functional/) doi:10.1109/CVPR.2011.5995582

BibTeX

@inproceedings{gall2011cvpr-functional,
  title     = {{Functional Categorization of Objects Using Real-Time Markerless Motion Capture}},
  author    = {Gall, Juergen and Fossati, Andrea and Van Gool, Luc},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  year      = {2011},
  pages     = {1969-1976},
  doi       = {10.1109/CVPR.2011.5995582},
  url       = {https://mlanthology.org/cvpr/2011/gall2011cvpr-functional/}
}