Generalized Zero-Shot Learning via Disentangled Representation
Abstract
Zero-Shot Learning (ZSL) aims to recognize images belonging to unseen classes that are unavailable in the training process, while Generalized Zero-Shot Learning (GZSL) is a more realistic variant that both seen and unseen classes appear during testing. Most GZSL approaches achieve knowledge transfer based on the features of samples that inevitably contain information irrelevant to recognition, bringing negative influence for the performance. In this work, we propose a novel method, dubbed Disentangled-VAE, which aims to disentangle category-distilling factors and category-dispersing factors from visual as well as semantic features, respectively. In addition, a batch re-combining strategy on latent features is introduced to guide the disentanglement, encouraging the distilling latent features to be more discriminative for recognition. Extensive experiments demonstrate that our method outperforms the state-of-the-art approaches on four challenging benchmark datasets
Cite
Text
Li et al. "Generalized Zero-Shot Learning via Disentangled Representation." AAAI Conference on Artificial Intelligence, 2021. doi:10.1609/AAAI.V35I3.16292Markdown
[Li et al. "Generalized Zero-Shot Learning via Disentangled Representation." AAAI Conference on Artificial Intelligence, 2021.](https://mlanthology.org/aaai/2021/li2021aaai-generalized/) doi:10.1609/AAAI.V35I3.16292BibTeX
@inproceedings{li2021aaai-generalized,
title = {{Generalized Zero-Shot Learning via Disentangled Representation}},
author = {Li, Xiangyu and Xu, Zhe and Wei, Kun and Deng, Cheng},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2021},
pages = {1966-1974},
doi = {10.1609/AAAI.V35I3.16292},
url = {https://mlanthology.org/aaai/2021/li2021aaai-generalized/}
}