A Novel Adversarial Inference Framework for Video Prediction with Action Control
Abstract
The ability of predicting future frames in video sequences, known as video prediction, is an appealing yet challenging task in computer vision. This task requires an in-depth representation of video sequences and a deep understanding of real-word causal rules. Existing approaches often result in blur predictions and lack the ability of action control. To tackle these problems, we propose a framework, called VPGAN, which employs an adversarial inference model and a cycle-consistency loss function to empower the framework to obtain more accurate predictions. In addition, we incorporate a conformal mapping network structure into VPGAN to enable action control for generating desirable future frames. In this way, VPGAN is able to produce fake videos of an object moving along a specific direction. Experimental results show that a combination of VPGAN with some pre-trained image segmentation models outperforms existing stochastic video prediction methods.
Cite
Text
Hu and Wang. "A Novel Adversarial Inference Framework for Video Prediction with Action Control." IEEE/CVF International Conference on Computer Vision Workshops, 2019. doi:10.1109/ICCVW.2019.00101Markdown
[Hu and Wang. "A Novel Adversarial Inference Framework for Video Prediction with Action Control." IEEE/CVF International Conference on Computer Vision Workshops, 2019.](https://mlanthology.org/iccvw/2019/hu2019iccvw-novel/) doi:10.1109/ICCVW.2019.00101BibTeX
@inproceedings{hu2019iccvw-novel,
title = {{A Novel Adversarial Inference Framework for Video Prediction with Action Control}},
author = {Hu, Zhihang and Wang, Jason Tsong-Li},
booktitle = {IEEE/CVF International Conference on Computer Vision Workshops},
year = {2019},
pages = {768-772},
doi = {10.1109/ICCVW.2019.00101},
url = {https://mlanthology.org/iccvw/2019/hu2019iccvw-novel/}
}