Unpaired Faces to Cartoons: Improving XGAN

Abstract

Domain Adaptation is a task that aims to translate an image from a source domain to a desired target domain. Current methods in domain adaptation use adversarial training based on Generative Adversarial Networks (GAN). In the present work, we focus on the task of domain adaptation from real faces to cartoon face images. We start from a baseline architecture called XGAN and introduce some improvements to it. Our proposed model is called W-XDGAN, which uses a form of GAN called Wasserstein-GAN, learns to approximate the Wasserstein Distance, and adds a denoiser to smooth the output cartoons. Whereas the original XGAN paper only presented a qualitative analysis, the advantages of this solution are demonstrated both quantitatively and qualitatively by comparing the results with models such as UNIT and original XGAN. Our code and models are publicly available at https://github.com/IAmigos/avatar-image-generator.

Cite

Text

Ramos et al. "Unpaired Faces to Cartoons: Improving XGAN." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2022. doi:10.1109/CVPRW56347.2022.00158

Markdown

[Ramos et al. "Unpaired Faces to Cartoons: Improving XGAN." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2022.](https://mlanthology.org/cvprw/2022/ramos2022cvprw-unpaired/) doi:10.1109/CVPRW56347.2022.00158

BibTeX

@inproceedings{ramos2022cvprw-unpaired,
  title     = {{Unpaired Faces to Cartoons: Improving XGAN}},
  author    = {Ramos, Stev H. and Cabrera, Joel and Ibáñez, Daniel and Jiménez-Panta, Alejandro B. and Castañón, César Beltrán and Villanueva, Edwin},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2022},
  pages     = {1517-1526},
  doi       = {10.1109/CVPRW56347.2022.00158},
  url       = {https://mlanthology.org/cvprw/2022/ramos2022cvprw-unpaired/}
}