Temporal Action Localization by Structured Maximal Sums
Abstract
We address the problem of temporal action localization in videos. We pose action localization as a structured prediction over arbitrary-length temporal windows, where each window is scored as the sum of frame-wise classification scores. Additionally, our model classifies the start, middle, and end of each action as separate components, allowing our system to explicitly model each action's temporal evolution and take advantage of informative temporal dependencies present in that structure. In this framework, we localize actions by searching for the structured maximal sum, a problem for which we develop a novel, provably-efficient algorithmic solution. The frame-wise classification scores are computed using features from a deep Convolutional Neural Network (CNN), which are trained end-to-end to directly optimize for a novel structured objective. We evaluate our system on the THUMOS '14 action detection benchmark and achieve competitive performance.
Cite
Text
Yuan et al. "Temporal Action Localization by Structured Maximal Sums." Conference on Computer Vision and Pattern Recognition, 2017. doi:10.1109/CVPR.2017.342Markdown
[Yuan et al. "Temporal Action Localization by Structured Maximal Sums." Conference on Computer Vision and Pattern Recognition, 2017.](https://mlanthology.org/cvpr/2017/yuan2017cvpr-temporal/) doi:10.1109/CVPR.2017.342BibTeX
@inproceedings{yuan2017cvpr-temporal,
title = {{Temporal Action Localization by Structured Maximal Sums}},
author = {Yuan, Zehuan and Stroud, Jonathan C. and Lu, Tong and Deng, Jia},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2017},
doi = {10.1109/CVPR.2017.342},
url = {https://mlanthology.org/cvpr/2017/yuan2017cvpr-temporal/}
}