Object-Based Illumination Estimation with Rendering-Aware Neural Networks
Abstract
We present a scheme for fast environment light estimation from the RGBD appearance of individual objects and their local image areas. Conventional inverse rendering is too computationally demanding for real-time applications, and the performance of purely learning-based techniques may be limited by the meager input data available from individual objects. To address these issues, we propose an approach that takes advantage of physical principles from inverse rendering to constrain the solution, while also utilizing neural networks to expedite the more computationally expensive portions of its processing, to increase robustness to noisy input data as well as to improve temporal and spatial stability. This results in a rendering-aware system that estimates the local illumination distribution at an object with high accuracy and in real time. With the estimated lighting, virtual objects can be rendered in AR scenarios with shading that is consistent to the real scene, leading to improved realism.
Cite
Text
Wei et al. "Object-Based Illumination Estimation with Rendering-Aware Neural Networks." Proceedings of the European Conference on Computer Vision (ECCV), 2020. doi:10.1007/978-3-030-58555-6_23Markdown
[Wei et al. "Object-Based Illumination Estimation with Rendering-Aware Neural Networks." Proceedings of the European Conference on Computer Vision (ECCV), 2020.](https://mlanthology.org/eccv/2020/wei2020eccv-objectbased/) doi:10.1007/978-3-030-58555-6_23BibTeX
@inproceedings{wei2020eccv-objectbased,
title = {{Object-Based Illumination Estimation with Rendering-Aware Neural Networks}},
author = {Wei, Xin and Chen, Guojun and Dong, Yue and Lin, Stephen and Tong, Xin},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2020},
doi = {10.1007/978-3-030-58555-6_23},
url = {https://mlanthology.org/eccv/2020/wei2020eccv-objectbased/}
}