Overlapping Clustering Models, and One (class) SVM to Bind Them All

Abstract

People belong to multiple communities, words belong to multiple topics, and books cover multiple genres; overlapping clusters are commonplace. Many existing overlapping clustering methods model each person (or word, or book) as a non-negative weighted combination of "exemplars" who belong solely to one community, with some small noise. Geometrically, each person is a point on a cone whose corners are these exemplars. This basic form encompasses the widely used Mixed Membership Stochastic Blockmodel of networks and its degree-corrected variants, as well as topic models such as LDA. We show that a simple one-class SVM yields provably consistent parameter inference for all such models, and scales to large datasets. Experimental results on several simulated and real datasets show our algorithm (called SVM-cone) is both accurate and scalable.

Cite

Text

Mao et al. "Overlapping Clustering Models, and One (class) SVM to Bind Them All." Neural Information Processing Systems, 2018.

Markdown

[Mao et al. "Overlapping Clustering Models, and One (class) SVM to Bind Them All." Neural Information Processing Systems, 2018.](https://mlanthology.org/neurips/2018/mao2018neurips-overlapping/)

BibTeX

@inproceedings{mao2018neurips-overlapping,
  title     = {{Overlapping Clustering Models, and One (class) SVM to Bind Them All}},
  author    = {Mao, Xueyu and Sarkar, Purnamrita and Chakrabarti, Deepayan},
  booktitle = {Neural Information Processing Systems},
  year      = {2018},
  pages     = {2126-2136},
  url       = {https://mlanthology.org/neurips/2018/mao2018neurips-overlapping/}
}