Learning Switchable Representation with Masked Decoding and Sparse Encoding
Abstract
In this study, we explore the unsupervised learning based on private/shared factor decomposition, which decomposes the latent space into private factors that vary only in a specific domain the shared factors that vary in all domains. We study when/how we can force the model to respect the true private/shared factor decomposition that underlies the dataset. We show that, when we train a masked decoder and an encoder with sparseness regularization in the latent space, we can identify the true private/shared decomposition up to mixing within each component. We empirically confirm this result and study the efficacy of this training strategy as a representation learning method.
Cite
Text
Hayashi and Koyama. "Learning Switchable Representation with Masked Decoding and Sparse Encoding." ICML 2022 Workshops: SCIS, 2022.Markdown
[Hayashi and Koyama. "Learning Switchable Representation with Masked Decoding and Sparse Encoding." ICML 2022 Workshops: SCIS, 2022.](https://mlanthology.org/icmlw/2022/hayashi2022icmlw-learning/)BibTeX
@inproceedings{hayashi2022icmlw-learning,
title = {{Learning Switchable Representation with Masked Decoding and Sparse Encoding}},
author = {Hayashi, Kohei and Koyama, Masanori},
booktitle = {ICML 2022 Workshops: SCIS},
year = {2022},
url = {https://mlanthology.org/icmlw/2022/hayashi2022icmlw-learning/}
}