Revisiting Score Function Estimators for $k$-Subset Sampling
Abstract
Are score function estimators an underestimated approach to learning with $k$-subset sampling? Sampling $k$-subsets is a fundamental operation in many machine learning tasks that is not amenable to differentiable parametrization, impeding gradient-based optimization. Prior work has focused on relaxed sampling or pathwise gradient estimators. Inspired by the success of score function estimators in variational inference and reinforcement learning, we revisit them within the context of $k$-subset sampling. Specifically, we demonstrate how to efficiently compute the $k$-subset distribution's score function using a discrete Fourier transform, and reduce the estimator's variance with control variates. The resulting estimator provides both exact samples and unbiased gradient estimates while also applying to non-differentiable downstream models, unlike existing methods. Experiments in feature selection show results competitive with current methods, despite weaker assumptions.
Cite
Text
Wijk et al. "Revisiting Score Function Estimators for $k$-Subset Sampling." ICML 2024 Workshops: Differentiable_Almost_Everything, 2024.Markdown
[Wijk et al. "Revisiting Score Function Estimators for $k$-Subset Sampling." ICML 2024 Workshops: Differentiable_Almost_Everything, 2024.](https://mlanthology.org/icmlw/2024/wijk2024icmlw-revisiting/)BibTeX
@inproceedings{wijk2024icmlw-revisiting,
title = {{Revisiting Score Function Estimators for $k$-Subset Sampling}},
author = {Wijk, Klas and Motilva, Ricardo Vinuesa and Azizpour, Hossein},
booktitle = {ICML 2024 Workshops: Differentiable_Almost_Everything},
year = {2024},
url = {https://mlanthology.org/icmlw/2024/wijk2024icmlw-revisiting/}
}