DYAN: A Dynamical Atoms-Based Network for Video Prediction

Abstract

The ability to anticipate the future is essential when making real time critical decisions, provides valuable information to understand dynamic natural scenes, and can help unsupervised video representation learning. State-of-art video prediction is based on complex architectures that need to learn large numbers of parameters, are potentially hard to train, slow to run, and may produce blurry predictions. In this paper, we introduce DYAN, a novel network with very few parameters and easy to train, which produces accurate, high quality frame predictions, significantly faster than previous approaches. DYAN owes its good qualities to its encoder and decoder, which are designed following concepts from systems identification theory and exploit the dynamics-based invariants of the data. Extensive experiments using several standard video datasets show that DYAN is superior generating frames and that it generalizes well across domains.

Cite

Text

Liu et al. "DYAN: A Dynamical Atoms-Based Network for Video Prediction." Proceedings of the European Conference on Computer Vision (ECCV), 2018. doi:10.1007/978-3-030-01258-8_11

Markdown

[Liu et al. "DYAN: A Dynamical Atoms-Based Network for Video Prediction." Proceedings of the European Conference on Computer Vision (ECCV), 2018.](https://mlanthology.org/eccv/2018/liu2018eccv-dyan/) doi:10.1007/978-3-030-01258-8_11

BibTeX

@inproceedings{liu2018eccv-dyan,
  title     = {{DYAN: A Dynamical Atoms-Based Network for Video Prediction}},
  author    = {Liu, Wenqian and Sharma, Abhishek and Camps, Octavia and Sznaier, Mario},
  booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
  year      = {2018},
  doi       = {10.1007/978-3-030-01258-8_11},
  url       = {https://mlanthology.org/eccv/2018/liu2018eccv-dyan/}
}