EDEN: Enhanced Diffusion for High-Quality Large-Motion Video Frame Interpolation
Abstract
Handling complex or nonlinear motion patterns has long posed challenges for video frame interpolation. Although recent advances in diffusion-based methods offer improvements over traditional optical flow-based approaches, they still struggle to generate sharp, temporally consistent frames in scenarios with large motion. To address this limitation, we introduce EDEN, an Enhanced Diffusion for high-quality large-motion vidEo frame iNterpolation. Our approach first utilizes a transformer-based tokenizer to produce refined latent representations of the intermediate frames for diffusion models. We then enhance the diffusion transformer with temporal attention across the process and incorporate a start-end frame difference embedding to guide the generation of dynamic motion. Extensive experiments demonstrate that EDEN achieves state-of-the-art results across popular benchmarks, including nearly a 10% LPIPS reduction on DAVIS and SNU-FILM, and an 8% improvement on DAIN-HD.
Cite
Text
Zhang et al. "EDEN: Enhanced Diffusion for High-Quality Large-Motion Video Frame Interpolation." Conference on Computer Vision and Pattern Recognition, 2025. doi:10.1109/CVPR52734.2025.00202Markdown
[Zhang et al. "EDEN: Enhanced Diffusion for High-Quality Large-Motion Video Frame Interpolation." Conference on Computer Vision and Pattern Recognition, 2025.](https://mlanthology.org/cvpr/2025/zhang2025cvpr-eden/) doi:10.1109/CVPR52734.2025.00202BibTeX
@inproceedings{zhang2025cvpr-eden,
title = {{EDEN: Enhanced Diffusion for High-Quality Large-Motion Video Frame Interpolation}},
author = {Zhang, Zihao and Chen, Haoran and Zhao, Haoyu and Lu, Guansong and Fu, Yanwei and Xu, Hang and Wu, Zuxuan},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2025},
pages = {2105-2115},
doi = {10.1109/CVPR52734.2025.00202},
url = {https://mlanthology.org/cvpr/2025/zhang2025cvpr-eden/}
}