Reparameterizable Subset Sampling via Continuous Relaxations
Abstract
Many machine learning tasks require sampling a subset of items from a collection based on a parameterized distribution. The Gumbel-softmax trick can be used to sample a single item, and allows for low-variance reparameterized gradients with respect to the parameters of the underlying distribution. However, stochastic optimization involving subset sampling is typically not reparameterizable. To overcome this limitation, we define a continuous relaxation of subset sampling that provides reparameterization gradients by generalizing the Gumbel-max trick. We use this approach to sample subsets of features in an instance-wise feature selection task for model interpretability, subsets of neighbors to implement a deep stochastic k-nearest neighbors model, and sub-sequences of neighbors to implement parametric t-SNE by directly comparing the identities of local neighbors. We improve performance in all these tasks by incorporating subset sampling in end-to-end training.
Cite
Text
Xie and Ermon. "Reparameterizable Subset Sampling via Continuous Relaxations." International Joint Conference on Artificial Intelligence, 2019. doi:10.24963/IJCAI.2019/544Markdown
[Xie and Ermon. "Reparameterizable Subset Sampling via Continuous Relaxations." International Joint Conference on Artificial Intelligence, 2019.](https://mlanthology.org/ijcai/2019/xie2019ijcai-reparameterizable/) doi:10.24963/IJCAI.2019/544BibTeX
@inproceedings{xie2019ijcai-reparameterizable,
title = {{Reparameterizable Subset Sampling via Continuous Relaxations}},
author = {Xie, Sang Michael and Ermon, Stefano},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2019},
pages = {3919-3925},
doi = {10.24963/IJCAI.2019/544},
url = {https://mlanthology.org/ijcai/2019/xie2019ijcai-reparameterizable/}
}