Neural Multi-View Self-Calibrated Photometric Stereo Without Photometric Stereo Cues
Abstract
We propose a neural inverse rendering approach that jointly reconstructs geometry, spatially varying reflectance, and lighting conditions from multi-view images captured under varying directional lighting. Unlike prior multi-view photometric stereo methods that require light calibration or intermediate cues such as per-view normal maps, our method jointly optimizes all scene parameters from raw images in a single stage. We represent both geometry and reflectance as neural implicit fields and apply shadow-aware volume rendering. A spatial network first predicts the signed distance and a reflectance latent code for each scene point. A reflectance network then estimates reflectance values conditioned on the latent code and angularly encoded surface normal, view, and light directions. The proposed method outperforms state-of-the-art normal-guided approaches in shape and lighting estimation accuracy, generalizes to view-unaligned multi-light images, and handles objects with challenging geometry and reflectance.
Cite
Text
Cao and Taketomi. "Neural Multi-View Self-Calibrated Photometric Stereo Without Photometric Stereo Cues." International Conference on Computer Vision, 2025.Markdown
[Cao and Taketomi. "Neural Multi-View Self-Calibrated Photometric Stereo Without Photometric Stereo Cues." International Conference on Computer Vision, 2025.](https://mlanthology.org/iccv/2025/cao2025iccv-neural/)BibTeX
@inproceedings{cao2025iccv-neural,
title = {{Neural Multi-View Self-Calibrated Photometric Stereo Without Photometric Stereo Cues}},
author = {Cao, Xu and Taketomi, Takafumi},
booktitle = {International Conference on Computer Vision},
year = {2025},
pages = {27552-27562},
url = {https://mlanthology.org/iccv/2025/cao2025iccv-neural/}
}