Scene Intrinsics and Depth from a Single Image

Abstract

Intrinsic image decomposition factorizes an observed image into its physical causes. This is most commonly framed as a decomposition into reflectance and shading, although recent progress has made full decompositions into shape, illumination, reflectance, and shading possible. However, existing factorization approaches require depth sensing to initialize the optimization of scene intrinsics. Rather than relying on depth sensors, we show that depth estimated purely from monocular appearance can provide sufficient cues for intrinsic image analysis. Our full intrinsic pipeline regresses depth by a fully convolutional network then jointly optimizes the intrinsic factorization to recover the input image. This combination yields full decompositions by uniting feature learning through deep network regression with physical modeling through statistical priors and random field regularization. This work demonstrates the first pipeline for full intrinsic decomposition of scenes from a single color image input alone.

Cite

Text

Shelhamer et al. "Scene Intrinsics and Depth from a Single Image." IEEE/CVF International Conference on Computer Vision Workshops, 2015. doi:10.1109/ICCVW.2015.39

Markdown

[Shelhamer et al. "Scene Intrinsics and Depth from a Single Image." IEEE/CVF International Conference on Computer Vision Workshops, 2015.](https://mlanthology.org/iccvw/2015/shelhamer2015iccvw-scene/) doi:10.1109/ICCVW.2015.39

BibTeX

@inproceedings{shelhamer2015iccvw-scene,
  title     = {{Scene Intrinsics and Depth from a Single Image}},
  author    = {Shelhamer, Evan and Barron, Jonathan T. and Darrell, Trevor},
  booktitle = {IEEE/CVF International Conference on Computer Vision Workshops},
  year      = {2015},
  pages     = {235-242},
  doi       = {10.1109/ICCVW.2015.39},
  url       = {https://mlanthology.org/iccvw/2015/shelhamer2015iccvw-scene/}
}