Material Editing Using a Physically Based Rendering Network

Abstract

The ability to edit materials of objects in images is desirable by many content creators. However, this is an extremely challenging task as it requires to disentangle intrinsic physical properties of an image. We propose an end-to-end network architecture that replicates the forward image formation process to accomplish this task. Specifically, given a single image, the network first predicts intrinsic properties, i.e. shape, illumination, and material, which are then provided to a rendering layer. This layer performs in-network image synthesis, thereby enabling the network to understand the physics behind the image formation process. The proposed rendering layer is fully differentiable, supports both diffuse and specular materials, and thus can be applicable in a variety of problem settings. We demonstrate a rich set of visually plausible material editing examples and provide an extensive comparative study.

Cite

Text

Liu et al. "Material Editing Using a Physically Based Rendering Network." International Conference on Computer Vision, 2017. doi:10.1109/ICCV.2017.248

Markdown

[Liu et al. "Material Editing Using a Physically Based Rendering Network." International Conference on Computer Vision, 2017.](https://mlanthology.org/iccv/2017/liu2017iccv-material/) doi:10.1109/ICCV.2017.248

BibTeX

@inproceedings{liu2017iccv-material,
  title     = {{Material Editing Using a Physically Based Rendering Network}},
  author    = {Liu, Guilin and Ceylan, Duygu and Yumer, Ersin and Yang, Jimei and Lien, Jyh-Ming},
  booktitle = {International Conference on Computer Vision},
  year      = {2017},
  doi       = {10.1109/ICCV.2017.248},
  url       = {https://mlanthology.org/iccv/2017/liu2017iccv-material/}
}