Tunable Dual-Objective GANs for Stable Training

Abstract

In an effort to address the training instabilities of GANs, we introduce a class of dual-objective GANs with different value functions (objectives) for the generator (G) and discriminator (D). In particular, we model each objective using $\alpha$-loss, a tunable classification loss, to obtain $(\alpha_D,\alpha_G)$-GANs, parameterized by $(\alpha_D,\alpha_G)\in (0,\infty]^2$. For sufficiently large number of samples and capacities for G and D, we show that the resulting non-zero sum game simplifies to minimizing an $f$-divergence under appropriate conditions on $(\alpha_D,\alpha_G)$. We highlight the value of tuning $(\alpha_D,\alpha_G)$ in alleviating training instabilities for the synthetic 2D Gaussian mixture ring, the Celeb-A, and the LSUN Classroom datasets.

Cite

Text

Welfert et al. "Tunable Dual-Objective GANs for Stable Training." ICML 2023 Workshops: AdvML-Frontiers, 2023.

Markdown

[Welfert et al. "Tunable Dual-Objective GANs for Stable Training." ICML 2023 Workshops: AdvML-Frontiers, 2023.](https://mlanthology.org/icmlw/2023/welfert2023icmlw-tunable/)

BibTeX

@inproceedings{welfert2023icmlw-tunable,
  title     = {{Tunable Dual-Objective GANs for Stable Training}},
  author    = {Welfert, Monica and Otstot, Kyle and Kurri, Gowtham Raghunath and Sankar, Lalitha},
  booktitle = {ICML 2023 Workshops: AdvML-Frontiers},
  year      = {2023},
  url       = {https://mlanthology.org/icmlw/2023/welfert2023icmlw-tunable/}
}