IMFine: 3D Inpainting via Geometry-Guided Multi-View Refinement

Abstract

Current 3D inpainting and object removal methods are largely limited to front-facing scenes, facing substantial challenges when applied to diverse, "unconstrained" scenes where the camera orientation and trajectory are unrestricted. To bridge this gap, we introduce a novel approach that produces inpainted 3D scenes with consistent visual quality and coherent underlying geometry across both front-facing and unconstrained scenes. Specifically, we propose a robust 3D inpainting pipeline that incorporates geometric priors and a multi-view refinement network trained via test-time adaptation, building on a pre-trained image inpainting model. Additionally, we develop a novel inpainting mask detection technique to derive targeted inpainting masks from object masks, boosting the performance in handling unconstrained scenes. To validate the efficacy of our approach, we create a challenging and diverse benchmark that spans a wide range of scenes. Comprehensive experiments demonstrate that our proposed method substantially outperforms existing state-of-the-art approaches.

Cite

Text

Shi et al. "IMFine: 3D Inpainting via Geometry-Guided Multi-View Refinement." Conference on Computer Vision and Pattern Recognition, 2025. doi:10.1109/CVPR52734.2025.02486

Markdown

[Shi et al. "IMFine: 3D Inpainting via Geometry-Guided Multi-View Refinement." Conference on Computer Vision and Pattern Recognition, 2025.](https://mlanthology.org/cvpr/2025/shi2025cvpr-imfine/) doi:10.1109/CVPR52734.2025.02486

BibTeX

@inproceedings{shi2025cvpr-imfine,
  title     = {{IMFine: 3D Inpainting via Geometry-Guided Multi-View Refinement}},
  author    = {Shi, Zhihao and Huo, Dong and Zhou, Yuhongze and Min, Yan and Lu, Juwei and Zuo, Xinxin},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2025},
  pages     = {26694-26703},
  doi       = {10.1109/CVPR52734.2025.02486},
  url       = {https://mlanthology.org/cvpr/2025/shi2025cvpr-imfine/}
}