SEINE: Short-to-Long Video Diffusion Model for Generative Transition and Prediction

Abstract

Recently video generation has achieved substantial progress with realistic results. Nevertheless, existing AI-generated videos are usually very short clips ("shot-level'') depicting a single scene. To deliver a coherent long video ("story-level''), it is desirable to have creative transition and prediction effects across different clips. This paper presents a short-to-long video diffusion model, SEINE, that focuses on generative transition and prediction. The goal is to generate high-quality long videos with smooth and creative transitions between scenes and varying lengths of shot-level videos. Specifically, we propose a random-mask video diffusion model to automatically generate transitions based on textual descriptions. By providing the images of different scenes as inputs, combined with text-based control, our model generates transition videos that ensure coherence and visual quality. Furthermore, the model can be readily extended to various tasks such as image-to-video animation and autoregressive video prediction. To conduct a comprehensive evaluation of this new generative task, we propose three assessing criteria for smooth and creative transition: temporal consistency, semantic similarity, and video-text semantic alignment. Extensive experiments validate the effectiveness of our approach over existing methods for generative transition and prediction, enabling the creation of story-level long videos.

Cite

Text

Chen et al. "SEINE: Short-to-Long Video Diffusion Model for Generative Transition and Prediction." International Conference on Learning Representations, 2024.

Markdown

[Chen et al. "SEINE: Short-to-Long Video Diffusion Model for Generative Transition and Prediction." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/chen2024iclr-seine/)

BibTeX

@inproceedings{chen2024iclr-seine,
  title     = {{SEINE: Short-to-Long Video Diffusion Model for Generative Transition and Prediction}},
  author    = {Chen, Xinyuan and Wang, Yaohui and Zhang, Lingjun and Zhuang, Shaobin and Ma, Xin and Yu, Jiashuo and Wang, Yali and Lin, Dahua and Qiao, Yu and Liu, Ziwei},
  booktitle = {International Conference on Learning Representations},
  year      = {2024},
  url       = {https://mlanthology.org/iclr/2024/chen2024iclr-seine/}
}