Zero-Shot Learning via Class-Conditioned Deep Generative Models

Abstract

We present a deep generative model for Zero-Shot Learning (ZSL). Unlike most existing methods for this problem, that represent each class as a point (via a semantic embedding), we represent each seen/unseen class using a class-specific latent-space distribution, conditioned on class attributes. We use these latent-space distributions as a prior for a supervised variational autoencoder (VAE), which also facilitates learning highly discriminative feature representations for the inputs. The entire framework is learned end-to-end using only the seen-class training data. At test time, the label for an unseen-class test input is the class that maximizes the VAE lower bound. We further extend the model to a (i) semi-supervised/transductive setting by leveraging unlabeled unseen-class data via an unsupervised learning module, and (ii) few-shot learning where we also have a small number of labeled inputs from the unseen classes. We compare our model with several state-of-the-art methods through a comprehensive set of experiments on a variety of benchmark data sets.

Cite

Text

Wang et al. "Zero-Shot Learning via Class-Conditioned Deep Generative Models." AAAI Conference on Artificial Intelligence, 2018. doi:10.1609/AAAI.V32I1.11600

Markdown

[Wang et al. "Zero-Shot Learning via Class-Conditioned Deep Generative Models." AAAI Conference on Artificial Intelligence, 2018.](https://mlanthology.org/aaai/2018/wang2018aaai-zero/) doi:10.1609/AAAI.V32I1.11600

BibTeX

@inproceedings{wang2018aaai-zero,
  title     = {{Zero-Shot Learning via Class-Conditioned Deep Generative Models}},
  author    = {Wang, Wenlin and Pu, Yunchen and Verma, Vinay Kumar and Fan, Kai and Zhang, Yizhe and Chen, Changyou and Rai, Piyush and Carin, Lawrence},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2018},
  pages     = {4211-4218},
  doi       = {10.1609/AAAI.V32I1.11600},
  url       = {https://mlanthology.org/aaai/2018/wang2018aaai-zero/}
}