IMFine: 3D Inpainting via Geometry-Guided Multi-View Refinement
Abstract
Current 3D inpainting and object removal methods are largely limited to front-facing scenes, facing substantial challenges when applied to diverse, "unconstrained" scenes where the camera orientation and trajectory are unrestricted. To bridge this gap, we introduce a novel approach that produces inpainted 3D scenes with consistent visual quality and coherent underlying geometry across both front-facing and unconstrained scenes. Specifically, we propose a robust 3D inpainting pipeline that incorporates geometric priors and a multi-view refinement network trained via test-time adaptation, building on a pre-trained image inpainting model. Additionally, we develop a novel inpainting mask detection technique to derive targeted inpainting masks from object masks, boosting the performance in handling unconstrained scenes. To validate the efficacy of our approach, we create a challenging and diverse benchmark that spans a wide range of scenes. Comprehensive experiments demonstrate that our proposed method substantially outperforms existing state-of-the-art approaches.
Cite
Text
Shi et al. "IMFine: 3D Inpainting via Geometry-Guided Multi-View Refinement." Conference on Computer Vision and Pattern Recognition, 2025. doi:10.1109/CVPR52734.2025.02486Markdown
[Shi et al. "IMFine: 3D Inpainting via Geometry-Guided Multi-View Refinement." Conference on Computer Vision and Pattern Recognition, 2025.](https://mlanthology.org/cvpr/2025/shi2025cvpr-imfine/) doi:10.1109/CVPR52734.2025.02486BibTeX
@inproceedings{shi2025cvpr-imfine,
title = {{IMFine: 3D Inpainting via Geometry-Guided Multi-View Refinement}},
author = {Shi, Zhihao and Huo, Dong and Zhou, Yuhongze and Min, Yan and Lu, Juwei and Zuo, Xinxin},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2025},
pages = {26694-26703},
doi = {10.1109/CVPR52734.2025.02486},
url = {https://mlanthology.org/cvpr/2025/shi2025cvpr-imfine/}
}