Learning to Optimize Combinatorial Functions
Abstract
Submodular functions have become a ubiquitous tool in machine learning. They are learnable from data, and can be optimized efficiently and with guarantees. Nonetheless, recent negative results show that optimizing learned surrogates of submodular functions can result in arbitrarily bad approximations of the true optimum. Our goal in this paper is to highlight the source of this hardness, and propose an alternative criterion for optimizing general combinatorial functions from sampled data. We prove a tight equivalence showing that a class of functions is optimizable if and only if it can be learned. We provide efficient and scalable optimization algorithms for several function classes of interest, and demonstrate their utility on the task of optimally choosing trending social media items.
Cite
Text
Rosenfeld et al. "Learning to Optimize Combinatorial Functions." International Conference on Machine Learning, 2018.Markdown
[Rosenfeld et al. "Learning to Optimize Combinatorial Functions." International Conference on Machine Learning, 2018.](https://mlanthology.org/icml/2018/rosenfeld2018icml-learning/)BibTeX
@inproceedings{rosenfeld2018icml-learning,
title = {{Learning to Optimize Combinatorial Functions}},
author = {Rosenfeld, Nir and Balkanski, Eric and Globerson, Amir and Singer, Yaron},
booktitle = {International Conference on Machine Learning},
year = {2018},
pages = {4374-4383},
volume = {80},
url = {https://mlanthology.org/icml/2018/rosenfeld2018icml-learning/}
}