EDEN: Enhanced Diffusion for High-Quality Large-Motion Video Frame Interpolation

Abstract

Handling complex or nonlinear motion patterns has long posed challenges for video frame interpolation. Although recent advances in diffusion-based methods offer improvements over traditional optical flow-based approaches, they still struggle to generate sharp, temporally consistent frames in scenarios with large motion. To address this limitation, we introduce EDEN, an Enhanced Diffusion for high-quality large-motion vidEo frame iNterpolation. Our approach first utilizes a transformer-based tokenizer to produce refined latent representations of the intermediate frames for diffusion models. We then enhance the diffusion transformer with temporal attention across the process and incorporate a start-end frame difference embedding to guide the generation of dynamic motion. Extensive experiments demonstrate that EDEN achieves state-of-the-art results across popular benchmarks, including nearly a 10% LPIPS reduction on DAVIS and SNU-FILM, and an 8% improvement on DAIN-HD.

Cite

Text

Zhang et al. "EDEN: Enhanced Diffusion for High-Quality Large-Motion Video Frame Interpolation." Conference on Computer Vision and Pattern Recognition, 2025. doi:10.1109/CVPR52734.2025.00202

Markdown

[Zhang et al. "EDEN: Enhanced Diffusion for High-Quality Large-Motion Video Frame Interpolation." Conference on Computer Vision and Pattern Recognition, 2025.](https://mlanthology.org/cvpr/2025/zhang2025cvpr-eden/) doi:10.1109/CVPR52734.2025.00202

BibTeX

@inproceedings{zhang2025cvpr-eden,
  title     = {{EDEN: Enhanced Diffusion for High-Quality Large-Motion Video Frame Interpolation}},
  author    = {Zhang, Zihao and Chen, Haoran and Zhao, Haoyu and Lu, Guansong and Fu, Yanwei and Xu, Hang and Wu, Zuxuan},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2025},
  pages     = {2105-2115},
  doi       = {10.1109/CVPR52734.2025.00202},
  url       = {https://mlanthology.org/cvpr/2025/zhang2025cvpr-eden/}
}