Reparameterizable Subset Sampling via Continuous Relaxations

Abstract

Many machine learning tasks require sampling a subset of items from a collection based on a parameterized distribution. The Gumbel-softmax trick can be used to sample a single item, and allows for low-variance reparameterized gradients with respect to the parameters of the underlying distribution. However, stochastic optimization involving subset sampling is typically not reparameterizable. To overcome this limitation, we define a continuous relaxation of subset sampling that provides reparameterization gradients by generalizing the Gumbel-max trick. We use this approach to sample subsets of features in an instance-wise feature selection task for model interpretability, subsets of neighbors to implement a deep stochastic k-nearest neighbors model, and sub-sequences of neighbors to implement parametric t-SNE by directly comparing the identities of local neighbors. We improve performance in all these tasks by incorporating subset sampling in end-to-end training.

Cite

Text

Xie and Ermon. "Reparameterizable Subset Sampling via Continuous Relaxations." International Joint Conference on Artificial Intelligence, 2019. doi:10.24963/IJCAI.2019/544

Markdown

[Xie and Ermon. "Reparameterizable Subset Sampling via Continuous Relaxations." International Joint Conference on Artificial Intelligence, 2019.](https://mlanthology.org/ijcai/2019/xie2019ijcai-reparameterizable/) doi:10.24963/IJCAI.2019/544

BibTeX

@inproceedings{xie2019ijcai-reparameterizable,
  title     = {{Reparameterizable Subset Sampling via Continuous Relaxations}},
  author    = {Xie, Sang Michael and Ermon, Stefano},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2019},
  pages     = {3919-3925},
  doi       = {10.24963/IJCAI.2019/544},
  url       = {https://mlanthology.org/ijcai/2019/xie2019ijcai-reparameterizable/}
}