Blur Interpolation Transformer for Real-World Motion from Blur
Abstract
This paper studies the challenging problem of recovering motion from blur, also known as joint deblurring and interpolation or blur temporal super-resolution. The challenges are twofold: 1) the current methods still leave considerable room for improvement in terms of visual quality even on the synthetic dataset, and 2) poor generalization to real-world data. To this end, we propose a blur interpolation transformer (BiT) to effectively unravel the underlying temporal correlation encoded in blur. Based on multi-scale residual Swin transformer blocks, we introduce dual-end temporal supervision and temporally symmetric ensembling strategies to generate effective features for time-varying motion rendering. In addition, we design a hybrid camera system to collect the first real-world dataset of one-to-many blur-sharp video pairs. Experimental results show that BiT has a significant gain over the state-of-the-art methods on the public dataset Adobe240. Besides, the proposed real-world dataset effectively helps the model generalize well to real blurry scenarios. Code and data are available at https://github.com/zzh-tech/BiT.
Cite
Text
Zhong et al. "Blur Interpolation Transformer for Real-World Motion from Blur." Conference on Computer Vision and Pattern Recognition, 2023. doi:10.1109/CVPR52729.2023.00553Markdown
[Zhong et al. "Blur Interpolation Transformer for Real-World Motion from Blur." Conference on Computer Vision and Pattern Recognition, 2023.](https://mlanthology.org/cvpr/2023/zhong2023cvpr-blur/) doi:10.1109/CVPR52729.2023.00553BibTeX
@inproceedings{zhong2023cvpr-blur,
title = {{Blur Interpolation Transformer for Real-World Motion from Blur}},
author = {Zhong, Zhihang and Cao, Mingdeng and Ji, Xiang and Zheng, Yinqiang and Sato, Imari},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2023},
pages = {5713-5723},
doi = {10.1109/CVPR52729.2023.00553},
url = {https://mlanthology.org/cvpr/2023/zhong2023cvpr-blur/}
}