Dual-Domain Image Synthesis Using Segmentation-Guided GAN

Abstract

We introduce a segmentation-guided approach to synthesise images that integrate features from two distinct domains. Images synthesised by our dual-domain model belong to one domain within the semantic-mask, and to another in the rest of the image - smoothly integrated. We build on the successes of few-shot StyleGAN and single-shot semantic segmentation to minimise the amount of training required in utilising two domains.The method combines few-shot cross-domain StyleGAN with a latent optimiser to achieve images containing features of two distinct domains. We use a segmentation-guided perceptual loss, which compares both pixel-level and activations between domain-specific and dual-domain synthetic images. Results demonstrate qualitatively and quantitatively that our model is capable of synthesising dual-domain images on a variety of objects (faces, horses, cats, cars), domains (natural, caricature, sketches) and part-based masks (eyes, nose, mouth, hair, car bonnet). The code is publicly available1.

Cite

Text

Bazazian et al. "Dual-Domain Image Synthesis Using Segmentation-Guided GAN." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2022. doi:10.1109/CVPRW56347.2022.00066

Markdown

[Bazazian et al. "Dual-Domain Image Synthesis Using Segmentation-Guided GAN." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2022.](https://mlanthology.org/cvprw/2022/bazazian2022cvprw-dualdomain/) doi:10.1109/CVPRW56347.2022.00066

BibTeX

@inproceedings{bazazian2022cvprw-dualdomain,
  title     = {{Dual-Domain Image Synthesis Using Segmentation-Guided GAN}},
  author    = {Bazazian, Dena and Calway, Andrew and Damen, Dima},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2022},
  pages     = {506-515},
  doi       = {10.1109/CVPRW56347.2022.00066},
  url       = {https://mlanthology.org/cvprw/2022/bazazian2022cvprw-dualdomain/}
}