Neural Multi-View Self-Calibrated Photometric Stereo Without Photometric Stereo Cues

Abstract

We propose a neural inverse rendering approach that jointly reconstructs geometry, spatially varying reflectance, and lighting conditions from multi-view images captured under varying directional lighting. Unlike prior multi-view photometric stereo methods that require light calibration or intermediate cues such as per-view normal maps, our method jointly optimizes all scene parameters from raw images in a single stage. We represent both geometry and reflectance as neural implicit fields and apply shadow-aware volume rendering. A spatial network first predicts the signed distance and a reflectance latent code for each scene point. A reflectance network then estimates reflectance values conditioned on the latent code and angularly encoded surface normal, view, and light directions. The proposed method outperforms state-of-the-art normal-guided approaches in shape and lighting estimation accuracy, generalizes to view-unaligned multi-light images, and handles objects with challenging geometry and reflectance.

Cite

Text

Cao and Taketomi. "Neural Multi-View Self-Calibrated Photometric Stereo Without Photometric Stereo Cues." International Conference on Computer Vision, 2025.

Markdown

[Cao and Taketomi. "Neural Multi-View Self-Calibrated Photometric Stereo Without Photometric Stereo Cues." International Conference on Computer Vision, 2025.](https://mlanthology.org/iccv/2025/cao2025iccv-neural/)

BibTeX

@inproceedings{cao2025iccv-neural,
  title     = {{Neural Multi-View Self-Calibrated Photometric Stereo Without Photometric Stereo Cues}},
  author    = {Cao, Xu and Taketomi, Takafumi},
  booktitle = {International Conference on Computer Vision},
  year      = {2025},
  pages     = {27552-27562},
  url       = {https://mlanthology.org/iccv/2025/cao2025iccv-neural/}
}