VideoFlow: A Conditional Flow-Based Model for Stochastic Video Generation
Abstract
Generative models that can model and predict sequences of future events can, in principle, learn to capture complex real-world phenomena, such as physical interactions. However, a central challenge in video prediction is that the future is highly uncertain: a sequence of past observations of events can imply many possible futures. Although a number of recent works have studied probabilistic models that can represent uncertain futures, such models are either extremely expensive computationally as in the case of pixel-level autoregressive models, or do not directly optimize the likelihood of the data. To our knowledge, our work is the first to propose multi-frame video prediction with normalizing flows, which allows for direct optimization of the data likelihood, and produces high-quality stochastic predictions. We describe an approach for modeling the latent space dynamics, and demonstrate that flow-based generative models offer a viable and competitive approach to generative modeling of video.
Cite
Text
Kumar et al. "VideoFlow: A Conditional Flow-Based Model for Stochastic Video Generation." International Conference on Learning Representations, 2020.Markdown
[Kumar et al. "VideoFlow: A Conditional Flow-Based Model for Stochastic Video Generation." International Conference on Learning Representations, 2020.](https://mlanthology.org/iclr/2020/kumar2020iclr-videoflow/)BibTeX
@inproceedings{kumar2020iclr-videoflow,
title = {{VideoFlow: A Conditional Flow-Based Model for Stochastic Video Generation}},
author = {Kumar, Manoj and Babaeizadeh, Mohammad and Erhan, Dumitru and Finn, Chelsea and Levine, Sergey and Dinh, Laurent and Kingma, Durk},
booktitle = {International Conference on Learning Representations},
year = {2020},
url = {https://mlanthology.org/iclr/2020/kumar2020iclr-videoflow/}
}