Image Synthesis via Semantic Composition

Abstract

In this paper, we present a novel approach to synthesize realistic images based on their semantic layouts. It hypothesizes that for objects with similar appearance, they share similar representation. Our method establishes dependencies between regions according to their appearance correlation, yielding both spatially variant and associated representations. Conditioning on these features, we propose a dynamic weighted network constructed by spatially conditional computation (with both convolution and normalization). More than preserving semantic distinctions, the given dynamic network strengthens semantic relevance, benefiting global structure and detail synthesis. We demonstrate that our method gives the compelling generation performance qualitatively and quantitatively with extensive experiments on benchmarks.

Cite

Text

Wang et al. "Image Synthesis via Semantic Composition." International Conference on Computer Vision, 2021. doi:10.1109/ICCV48922.2021.01349

Markdown

[Wang et al. "Image Synthesis via Semantic Composition." International Conference on Computer Vision, 2021.](https://mlanthology.org/iccv/2021/wang2021iccv-image/) doi:10.1109/ICCV48922.2021.01349

BibTeX

@inproceedings{wang2021iccv-image,
  title     = {{Image Synthesis via Semantic Composition}},
  author    = {Wang, Yi and Qi, Lu and Chen, Ying-Cong and Zhang, Xiangyu and Jia, Jiaya},
  booktitle = {International Conference on Computer Vision},
  year      = {2021},
  pages     = {13749-13758},
  doi       = {10.1109/ICCV48922.2021.01349},
  url       = {https://mlanthology.org/iccv/2021/wang2021iccv-image/}
}