TR0N: Translator Networks for 0-Shot Plug-and-Play Conditional Generation
Abstract
We propose TR0N, a highly general framework to turn pre-trained unconditional generative models, such as GANs and VAEs, into conditional models. The conditioning can be highly arbitrary, and requires only a pre-trained auxiliary model. For example, we show how to turn unconditional models into class-conditional ones with the help of a classifier, and also into text-to-image models by leveraging CLIP. TR0N learns a lightweight stochastic mapping which "translates’" between the space of conditions and the latent space of the generative model, in such a way that the generated latent corresponds to a data sample satisfying the desired condition. The translated latent samples are then further improved upon through Langevin dynamics, enabling us to obtain higher-quality data samples. TR0N requires no training data nor fine-tuning, yet can achieve a zero-shot FID of 10.9 on MS-COCO, outperforming competing alternatives not only on this metric, but also in sampling speed – all while retaining a much higher level of generality. Our code is available at https://github.com/layer6ai-labs/tr0n.
Cite
Text
Liu et al. "TR0N: Translator Networks for 0-Shot Plug-and-Play Conditional Generation." International Conference on Machine Learning, 2023.Markdown
[Liu et al. "TR0N: Translator Networks for 0-Shot Plug-and-Play Conditional Generation." International Conference on Machine Learning, 2023.](https://mlanthology.org/icml/2023/liu2023icml-tr0n/)BibTeX
@inproceedings{liu2023icml-tr0n,
title = {{TR0N: Translator Networks for 0-Shot Plug-and-Play Conditional Generation}},
author = {Liu, Zhaoyan and Vouitsis, Noël and Gorti, Satya Krishna and Ba, Jimmy and Loaiza-Ganem, Gabriel},
booktitle = {International Conference on Machine Learning},
year = {2023},
pages = {22092-22112},
volume = {202},
url = {https://mlanthology.org/icml/2023/liu2023icml-tr0n/}
}