Deep Analysis of CNN-Based Spatio-Temporal Representations for Action Recognition
Abstract
In recent years, a number of approaches based on 2D or 3D convolutional neural networks (CNN) have emerged for video action recognition, achieving state-of-the-art results on several large-scale benchmark datasets. In this paper, we carry out in-depth comparative analysis to better understand the differences between these approaches and the progress made by them. To this end, we develop an unified framework for both 2D-CNN and 3D-CNN action models, which enables us to remove bells and whistles and provides a common ground for fair comparison. We then conduct an effort towards a large-scale analysis involving over 300 action recognition models. Our comprehensive analysis reveals that a) a significant leap is made in efficiency for action recognition, but not in accuracy; b) 2D-CNN and 3D-CNN models behave similarly in terms of spatio-temporal representation abilities and transferability. Our codes are available at https://github.com/IBM/action-recognition-pytorch.
Cite
Text
Chen et al. "Deep Analysis of CNN-Based Spatio-Temporal Representations for Action Recognition." Conference on Computer Vision and Pattern Recognition, 2021. doi:10.1109/CVPR46437.2021.00610Markdown
[Chen et al. "Deep Analysis of CNN-Based Spatio-Temporal Representations for Action Recognition." Conference on Computer Vision and Pattern Recognition, 2021.](https://mlanthology.org/cvpr/2021/chen2021cvpr-deep/) doi:10.1109/CVPR46437.2021.00610BibTeX
@inproceedings{chen2021cvpr-deep,
title = {{Deep Analysis of CNN-Based Spatio-Temporal Representations for Action Recognition}},
author = {Chen, Chun-Fu Richard and Panda, Rameswar and Ramakrishnan, Kandan and Feris, Rogerio and Cohn, John and Oliva, Aude and Fan, Quanfu},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2021},
pages = {6165-6175},
doi = {10.1109/CVPR46437.2021.00610},
url = {https://mlanthology.org/cvpr/2021/chen2021cvpr-deep/}
}