SIR-DIFF: Sparse Image Sets Restoration with Multi-View Diffusion Model
Abstract
The computer vision community has developed numerous techniques for digitally restoring true scene information from single-view degraded photographs, an important yet extremely ill-posed task. In this work, we tackle image restoration from a different perspective by jointly denoising multiple photographs of the same scene. Our core hypothesis is that degraded images capturing a shared scene contain complementary information that, when combined, better constrains the restoration problem. To this end, we implement a powerful multi-view diffusion model that jointly generates uncorrupted views by extracting rich information from multi-view relationships. Our experiments show that our multi-view approach outperforms existing single-view image and even video-based methods on image deblurring and super-resolution tasks. Critically, our model is trained to output 3D consistent images, making it a promising tool for applications requiring robust multi-view integration, such as 3D reconstruction or pose estimation.
Cite
Text
Mao et al. "SIR-DIFF: Sparse Image Sets Restoration with Multi-View Diffusion Model." Conference on Computer Vision and Pattern Recognition, 2025. doi:10.1109/CVPR52734.2025.02014Markdown
[Mao et al. "SIR-DIFF: Sparse Image Sets Restoration with Multi-View Diffusion Model." Conference on Computer Vision and Pattern Recognition, 2025.](https://mlanthology.org/cvpr/2025/mao2025cvpr-sirdiff/) doi:10.1109/CVPR52734.2025.02014BibTeX
@inproceedings{mao2025cvpr-sirdiff,
title = {{SIR-DIFF: Sparse Image Sets Restoration with Multi-View Diffusion Model}},
author = {Mao, Yucheng and Wang, Boyang and Kulkarni, Nilesh and Park, Jeong Joon},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2025},
pages = {21620-21630},
doi = {10.1109/CVPR52734.2025.02014},
url = {https://mlanthology.org/cvpr/2025/mao2025cvpr-sirdiff/}
}