ComboGAN: Unrestrained Scalability for Image Domain Translation

Abstract

The past year alone has seen unprecedented leaps in the area of learning-based image translation, namely Cycle-GAN, by Zhu et al. But experiments so far have been tailored to merely two domains at a time, and scaling them to more would require an quadratic number of models to be trained. And with two-domain models taking days to train on current hardware, the number of domains quickly becomes limited by the time and resources required to process them. In this paper, we propose a multi-component image translation model and training scheme which scales linearly - both in resource consumption and time required - with the number of domains. We demonstrate its capabilities on a dataset of paintings by 14 different artists and on images of the four different seasons in the Alps. Note that 14 data groups would need (14 choose 2) = 91 different CycleGAN models: a total of 182 generator/discriminator pairs; whereas our model requires only 14 generator/discriminator pairs.

Cite

Text

Anoosheh et al. "ComboGAN: Unrestrained Scalability for Image Domain Translation." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2018. doi:10.1109/CVPRW.2018.00122

Markdown

[Anoosheh et al. "ComboGAN: Unrestrained Scalability for Image Domain Translation." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2018.](https://mlanthology.org/cvprw/2018/anoosheh2018cvprw-combogan/) doi:10.1109/CVPRW.2018.00122

BibTeX

@inproceedings{anoosheh2018cvprw-combogan,
  title     = {{ComboGAN: Unrestrained Scalability for Image Domain Translation}},
  author    = {Anoosheh, Asha and Agustsson, Eirikur and Timofte, Radu and Van Gool, Luc},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2018},
  pages     = {783-790},
  doi       = {10.1109/CVPRW.2018.00122},
  url       = {https://mlanthology.org/cvprw/2018/anoosheh2018cvprw-combogan/}
}