Memory-Augmented Dense Predictive Coding for Video Representation Learning
Abstract
The objective of this paper is self-supervised learning from video, in particular for representations for action recognition. We make the following contributions: (i) We propose a new architecture and learning framework Memory-augmented Dense Predictive Coding (MemDPC) for the task. It is trained with a predictive attention mechanism over the set of compressed memories, such that any future states can always be constructed by a convex combination of the condense representations, allowing to make multiple hypotheses efficiently.(ii) We investigate visual-only self-supervised video representation learning from RGB frames, or from unsupervised optical flow, or both. (iii) We thoroughly evaluate the quality of learnt representation on four different downstream tasks: action recognition, video retrieval, learning with scarce annotations, and unintentional action classification. In all cases, we demonstrate state-of-the-art or comparable performance over other approaches with orders of magnitude fewer training data.
Cite
Text
Han et al. "Memory-Augmented Dense Predictive Coding for Video Representation Learning." Proceedings of the European Conference on Computer Vision (ECCV), 2020. doi:10.1007/978-3-030-58580-8_19Markdown
[Han et al. "Memory-Augmented Dense Predictive Coding for Video Representation Learning." Proceedings of the European Conference on Computer Vision (ECCV), 2020.](https://mlanthology.org/eccv/2020/han2020eccv-memoryaugmented/) doi:10.1007/978-3-030-58580-8_19BibTeX
@inproceedings{han2020eccv-memoryaugmented,
title = {{Memory-Augmented Dense Predictive Coding for Video Representation Learning}},
author = {Han, Tengda and Xie, Weidi and Zisserman, Andrew},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2020},
doi = {10.1007/978-3-030-58580-8_19},
url = {https://mlanthology.org/eccv/2020/han2020eccv-memoryaugmented/}
}