Intrinsic Video

Abstract

Intrinsic images such as albedo and shading are valuable for later stages of visual processing. Previous methods for extracting albedo and shading use either single images or images together with depth data. Instead, we define intrinsic video estimation as the problem of extracting temporally coherent albedo and shading from video alone. Our approach exploits the assumption that albedo is constant over time while shading changes slowly. Optical flow aids in the accurate estimation of intrinsic video by providing temporal continuity as well as putative surface boundaries. Additionally, we find that the estimated albedo sequence can be used to improve optical flow accuracy in sequences with changing illumination. The approach makes only weak assumptions about the scene and we show that it substantially outperforms existing single-frame intrinsic image methods. We evaluate this quantitatively on synthetic sequences as well on challenging natural sequences with complex geometry, motion, and illumination.

Cite

Text

Kong et al. "Intrinsic Video." European Conference on Computer Vision, 2014. doi:10.1007/978-3-319-10605-2_24

Markdown

[Kong et al. "Intrinsic Video." European Conference on Computer Vision, 2014.](https://mlanthology.org/eccv/2014/kong2014eccv-intrinsic/) doi:10.1007/978-3-319-10605-2_24

BibTeX

@inproceedings{kong2014eccv-intrinsic,
  title     = {{Intrinsic Video}},
  author    = {Kong, Naejin and Gehler, Peter V. and Black, Michael J.},
  booktitle = {European Conference on Computer Vision},
  year      = {2014},
  pages     = {360-375},
  doi       = {10.1007/978-3-319-10605-2_24},
  url       = {https://mlanthology.org/eccv/2014/kong2014eccv-intrinsic/}
}