Learning the Viewpoint Manifold for Action Recognition
Abstract
Researchers are increasingly interested in providing video-based, view-invariant action recognition for human motion. Addressing this problem will lead to more accurate modeling and analysis of the type of unconstrained video commonly collected in the areas of athletics and medicine. Previous viewpoint-invariant methods use multiple cameras in both the training and testing phases of action recognition or require storing many examples of a single action from multiple viewpoints. In this paper, we present a framework for learning a compact representation of primitive actions (e.g., walk, punch, kick, sit) that can be used for video obtained from a single camera for simultaneous action recognition and viewpoint estimation. Using our method, which models the low-dimensional structure of these actions relative to viewpoint, we show recognition rates on a publicly available data set previously only achieved using multiple simultaneous views.
Cite
Text
Souvenir and Babbs. "Learning the Viewpoint Manifold for Action Recognition." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2008. doi:10.1109/CVPR.2008.4587552Markdown
[Souvenir and Babbs. "Learning the Viewpoint Manifold for Action Recognition." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2008.](https://mlanthology.org/cvpr/2008/souvenir2008cvpr-learning/) doi:10.1109/CVPR.2008.4587552BibTeX
@inproceedings{souvenir2008cvpr-learning,
title = {{Learning the Viewpoint Manifold for Action Recognition}},
author = {Souvenir, Richard and Babbs, Justin},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition},
year = {2008},
doi = {10.1109/CVPR.2008.4587552},
url = {https://mlanthology.org/cvpr/2008/souvenir2008cvpr-learning/}
}