On Greedy Maximization of Entropy

Abstract

Submodular function maximization is one of the key problems that arise in many machine learning tasks. Greedy selection algorithms are the proven choice to solve such problems, where prior theoretical work guarantees (1 - 1/e) approximation ratio. However, it has been empirically observed that greedy selection provides almost optimal solutions in practice. The main goal of this paper is to explore and answer why the greedy selection does significantly better than the theoretical guarantee of (1 - 1/e). Applications include, but are not limited to, sensor selection tasks which use both entropy and mutual information as a maximization criteria. We give a theoretical justification for the nearly optimal approximation ratio via detailed analysis of the curvature of these objective functions for Gaussian RBF kernels.

Cite

Text

Sharma et al. "On Greedy Maximization of Entropy." International Conference on Machine Learning, 2015.

Markdown

[Sharma et al. "On Greedy Maximization of Entropy." International Conference on Machine Learning, 2015.](https://mlanthology.org/icml/2015/sharma2015icml-greedy/)

BibTeX

@inproceedings{sharma2015icml-greedy,
  title     = {{On Greedy Maximization of Entropy}},
  author    = {Sharma, Dravyansh and Kapoor, Ashish and Deshpande, Amit},
  booktitle = {International Conference on Machine Learning},
  year      = {2015},
  pages     = {1330-1338},
  volume    = {37},
  url       = {https://mlanthology.org/icml/2015/sharma2015icml-greedy/}
}