Mining Discriminative States of Hands and Objects to Recognize Egocentric Actions with a Wearable RGBD Camera

Abstract

Of increasing interest to the computer vision community is to recognize egocentric actions. Conceptually, an egocentric action is largely identifiable by the states of hands and objects. For example, "drinking soda" is essentially composed of two sequential states where one first "takes up the soda can", then "drinks from the soda can". While existing algorithms commonly use manually defined states to train action classifiers, we present a novel model that automatically mines discriminative states for recognizing egocentric actions. To mine discriminative states, we propose a novel kernel function and formulate a Multiple Kernel Learning based framework to learn adaptive weights for different states. Experiments on three benchmark datasets, i.e., RGBD-Ego, ADL, and GTEA, clearly show that our recognition algorithm outperforms state-of-the-art algorithms.

Cite

Text

Wan and Aggarwal. "Mining Discriminative States of Hands and Objects to Recognize Egocentric Actions with a Wearable RGBD Camera." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2015. doi:10.1109/CVPRW.2015.7301346

Markdown

[Wan and Aggarwal. "Mining Discriminative States of Hands and Objects to Recognize Egocentric Actions with a Wearable RGBD Camera." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2015.](https://mlanthology.org/cvprw/2015/wan2015cvprw-mining/) doi:10.1109/CVPRW.2015.7301346

BibTeX

@inproceedings{wan2015cvprw-mining,
  title     = {{Mining Discriminative States of Hands and Objects to Recognize Egocentric Actions with a Wearable RGBD Camera}},
  author    = {Wan, Shaohua and Aggarwal, J. K.},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2015},
  pages     = {36-43},
  doi       = {10.1109/CVPRW.2015.7301346},
  url       = {https://mlanthology.org/cvprw/2015/wan2015cvprw-mining/}
}