Learning a Discriminative Model for the Perception of Realism in Composite Images
Abstract
What makes an image appear realistic? In this work, we are answering this question from a data-driven perspective by learning the perception of visual realism directly from large amounts of data. In particular, we train a Convolutional Neural Network (CNN) model that distinguishes natural photographs from automatically generated composite images. The model learns to predict visual realism of a scene in terms of color, lighting and texture compatibility, without any human annotations pertaining to it. Our model outperforms previous works that rely on hand-crafted heuristics, for the task of classifying realistic vs. unrealistic photos. Furthermore, we apply our learned model to compute optimal parameters of a compositing method, to maximize the visual realism score predicted by our CNN model. We demonstrate its advantage against existing methods via a human perception study.
Cite
Text
Zhu et al. "Learning a Discriminative Model for the Perception of Realism in Composite Images." International Conference on Computer Vision, 2015. doi:10.1109/ICCV.2015.449Markdown
[Zhu et al. "Learning a Discriminative Model for the Perception of Realism in Composite Images." International Conference on Computer Vision, 2015.](https://mlanthology.org/iccv/2015/zhu2015iccv-learning/) doi:10.1109/ICCV.2015.449BibTeX
@inproceedings{zhu2015iccv-learning,
title = {{Learning a Discriminative Model for the Perception of Realism in Composite Images}},
author = {Zhu, Jun-Yan and Krahenbuhl, Philipp and Shechtman, Eli and Efros, Alexei A.},
booktitle = {International Conference on Computer Vision},
year = {2015},
doi = {10.1109/ICCV.2015.449},
url = {https://mlanthology.org/iccv/2015/zhu2015iccv-learning/}
}