Self-Supervised Motion Representation via Scattering Local Motion Cues
Abstract
Motion representation is key to many computer vision problems but has never been well studied in the literature. Existing works usually rely on the optical flow estimation to assist other tasks such as action recognition, frame prediction, video segmentation, etc. In this paper, we leverage the massive unlabeled video data to learn an accurate explicit motion representation that aligns well with the semantic distribution of the moving objects. Our method subsumes a coarse-to-fine paradigm, which first decodes the low-resolution motion maps from the rich spatial-temporal features of the video, then adaptively upsamples the low-resolution maps to the full-resolution by considering the semantic cues. To achieve this, we propose a novel context guided motion upsampling layer that leverages the spatial context of video objects to learn the upsampling parameters in an efficient way. We prove the effectiveness of our proposed motion representation method on downstream video understanding tasks, e.g., action recognition task. Experimental results show that our method performs favorably against state-of-the-art methods.
Cite
Text
Tian et al. "Self-Supervised Motion Representation via Scattering Local Motion Cues." Proceedings of the European Conference on Computer Vision (ECCV), 2020. doi:10.1007/978-3-030-58568-6_5Markdown
[Tian et al. "Self-Supervised Motion Representation via Scattering Local Motion Cues." Proceedings of the European Conference on Computer Vision (ECCV), 2020.](https://mlanthology.org/eccv/2020/tian2020eccv-selfsupervised/) doi:10.1007/978-3-030-58568-6_5BibTeX
@inproceedings{tian2020eccv-selfsupervised,
title = {{Self-Supervised Motion Representation via Scattering Local Motion Cues}},
author = {Tian, Yuan and Che, Zhaohui and Bao, Wenbo and Zhai, Guangtao and Gao, Zhiyong},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2020},
doi = {10.1007/978-3-030-58568-6_5},
url = {https://mlanthology.org/eccv/2020/tian2020eccv-selfsupervised/}
}