Learning Sparse Combinatorial Representations via Two-Stage Submodular Maximization

Abstract

We consider the problem of learning sparse representations of data sets, where the goal is to reduce a data set in manner that optimizes multiple objectives. Motivated by applications of data summarization, we develop a new model which we refer to as the two-stage submodular maximization problem. This task can be viewed as a combinatorial analogue of representation learning problems such as dictionary learning and sparse regression. The two-stage problem strictly generalizes the problem of cardinality constrained submodular maximization, though the objective function is not submodular and the techniques for submodular maximization cannot be applied. We describe a continuous optimization method which achieves an approximation ratio which asymptotically approaches 1-1/e. For instances where the asymptotics do not kick in, we design a local-search algorithm whose approximation ratio is arbitrarily close to 1/2. We empirically demonstrate the effectiveness of our methods on two multi-objective data summarization tasks, where the goal is to construct summaries via sparse representative subsets w.r.t. to predefined objectives.

Cite

Text

Balkanski et al. "Learning Sparse Combinatorial Representations via Two-Stage Submodular Maximization." International Conference on Machine Learning, 2016.

Markdown

[Balkanski et al. "Learning Sparse Combinatorial Representations via Two-Stage Submodular Maximization." International Conference on Machine Learning, 2016.](https://mlanthology.org/icml/2016/balkanski2016icml-learning/)

BibTeX

@inproceedings{balkanski2016icml-learning,
  title     = {{Learning Sparse Combinatorial Representations via Two-Stage Submodular Maximization}},
  author    = {Balkanski, Eric and Mirzasoleiman, Baharan and Krause, Andreas and Singer, Yaron},
  booktitle = {International Conference on Machine Learning},
  year      = {2016},
  pages     = {2207-2216},
  volume    = {48},
  url       = {https://mlanthology.org/icml/2016/balkanski2016icml-learning/}
}