Learning Procedure-Aware Video Representation from Instructional Videos and Their Narrations

Abstract

The abundance of instructional videos and their narrations over the Internet offers an exciting avenue for understanding procedural activities. In this work, we propose to learn video representation that encodes both action steps and their temporal ordering, based on a large-scale dataset of web instructional videos and their narrations, without using human annotations. Our method jointly learns a video representation to encode individual step concepts, and a deep probabilistic model to capture both temporal dependencies and immense individual variations in the step ordering. We empirically demonstrate that learning temporal ordering not only enables new capabilities for procedure reasoning, but also reinforces the recognition of individual steps. Our model significantly advances the state-of-the-art results on step classification (+2.8% / +3.3% on COIN / EPIC-Kitchens) and step forecasting (+7.4% on COIN). Moreover, our model attains promising results in zero-shot inference for step classification and forecasting, as well as in predicting diverse and plausible steps for incomplete procedures. Our code is available at https://github.com/facebookresearch/ProcedureVRL.

Cite

Text

Zhong et al. "Learning Procedure-Aware Video Representation from Instructional Videos and Their Narrations." Conference on Computer Vision and Pattern Recognition, 2023. doi:10.1109/CVPR52729.2023.01424

Markdown

[Zhong et al. "Learning Procedure-Aware Video Representation from Instructional Videos and Their Narrations." Conference on Computer Vision and Pattern Recognition, 2023.](https://mlanthology.org/cvpr/2023/zhong2023cvpr-learning/) doi:10.1109/CVPR52729.2023.01424

BibTeX

@inproceedings{zhong2023cvpr-learning,
  title     = {{Learning Procedure-Aware Video Representation from Instructional Videos and Their Narrations}},
  author    = {Zhong, Yiwu and Yu, Licheng and Bai, Yang and Li, Shangwen and Yan, Xueting and Li, Yin},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2023},
  pages     = {14825-14835},
  doi       = {10.1109/CVPR52729.2023.01424},
  url       = {https://mlanthology.org/cvpr/2023/zhong2023cvpr-learning/}
}