RC-MVSNet: Unsupervised Multi-View Stereo with Neural Rendering
Abstract
Finding accurate correspondences among different views is the Achilles’ heel of unsupervised Multi-View Stereo (MVS). Existing methods are built upon the assumption that corresponding pixels share similar photometric features. However, multi-view images in real scenarios observe non-Lambertian surfaces and experience occlusions. In this work, we propose a novel approach with neural rendering (RC-MVSNet) to solve such ambiguity issues of correspondences among views. Specifically, we impose a depth rendering consistency loss to constrain the geometry features close to the object surface to alleviate occlusions. Concurrently, we introduce a reference view synthesis loss to generate consistent supervision, even for non-Lambertian surfaces. Extensive experiments on DTU and Tanks&Temples benchmarks demonstrate that our RC-MVSNet approach achieves state-of-the-art performance over unsupervised MVS frameworks and competitive performance to many supervised methods. The code is released at https://github.com/Boese0601/RC-MVSNet.
Cite
Text
Chang et al. "RC-MVSNet: Unsupervised Multi-View Stereo with Neural Rendering." Proceedings of the European Conference on Computer Vision (ECCV), 2022. doi:10.1007/978-3-031-19821-2_38Markdown
[Chang et al. "RC-MVSNet: Unsupervised Multi-View Stereo with Neural Rendering." Proceedings of the European Conference on Computer Vision (ECCV), 2022.](https://mlanthology.org/eccv/2022/chang2022eccv-rcmvsnet/) doi:10.1007/978-3-031-19821-2_38BibTeX
@inproceedings{chang2022eccv-rcmvsnet,
title = {{RC-MVSNet: Unsupervised Multi-View Stereo with Neural Rendering}},
author = {Chang, Di and Božič, Aljaž and Zhang, Tong and Yan, Qingsong and Chen, Yingcong and Süsstrunk, Sabine and Nießner, Matthias},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2022},
doi = {10.1007/978-3-031-19821-2_38},
url = {https://mlanthology.org/eccv/2022/chang2022eccv-rcmvsnet/}
}