MVPSNet: Fast Generalizable Multi-View Photometric Stereo
Abstract
We propose a fast and generalizable solution to Multiview Photometric Stereo (MVPS), called MVPSNet. The key to our approach is a feature extraction network that effectively combines images from the same view captured under multiple lighting conditions to extract geometric features from shading cues for stereo matching. We demonstrate these features, termed 'Light Aggregated Feature Maps' (LAFM), are effective for feature matching even in textureless regions, where traditional multi-view stereo methods often fail. Our method produces similar reconstruction results to PS-NeRF, a state-of-the-art MVPS method that optimizes a neural network per-scene, while being 411x faster (105 seconds vs. 12 hours) in inference. Additionally, we introduce a new synthetic dataset for MVPS, sMVPS, which is shown to be effective for training a generalizable MVPS method.
Cite
Text
Zhao et al. "MVPSNet: Fast Generalizable Multi-View Photometric Stereo." International Conference on Computer Vision, 2023. doi:10.1109/ICCV51070.2023.01151Markdown
[Zhao et al. "MVPSNet: Fast Generalizable Multi-View Photometric Stereo." International Conference on Computer Vision, 2023.](https://mlanthology.org/iccv/2023/zhao2023iccv-mvpsnet/) doi:10.1109/ICCV51070.2023.01151BibTeX
@inproceedings{zhao2023iccv-mvpsnet,
title = {{MVPSNet: Fast Generalizable Multi-View Photometric Stereo}},
author = {Zhao, Dongxu and Lichy, Daniel and Perrin, Pierre-Nicolas and Frahm, Jan-Michael and Sengupta, Soumyadip},
booktitle = {International Conference on Computer Vision},
year = {2023},
pages = {12525-12536},
doi = {10.1109/ICCV51070.2023.01151},
url = {https://mlanthology.org/iccv/2023/zhao2023iccv-mvpsnet/}
}