Sparse Additive Text Models with Low Rank Background

Abstract

The sparse additive model for text modeling involves the sum-of-exp computing, with consuming costs for large scales. Moreover, the assumption of equal background across all classes/topics may be too strong. This paper extends to propose sparse additive model with low rank background (SAM-LRB), and simple yet efficient estimation. Particularly, by employing a double majorization bound, we approximate the log-likelihood into a quadratic lower-bound with the sum-of-exp terms absent. The constraints of low rank and sparsity are then simply embodied by nuclear norm and $\ell_1$-norm regularizers. Interestingly, we find that the optimization task in this manner can be transformed into the same form as that in Robust PCA. Consequently, parameters of supervised SAM-LRB can be efficiently learned using an existing algorithm for Robust PCA based on accelerated proximal gradient. Besides the supervised case, we extend SAM-LRB to also favor unsupervised and multifaceted scenarios. Experiments on real world data demonstrate the effectiveness and efficiency of SAM-LRB, showing state-of-the-art performances.

Cite

Text

Shi. "Sparse Additive Text Models with Low Rank Background." Neural Information Processing Systems, 2013.

Markdown

[Shi. "Sparse Additive Text Models with Low Rank Background." Neural Information Processing Systems, 2013.](https://mlanthology.org/neurips/2013/shi2013neurips-sparse/)

BibTeX

@inproceedings{shi2013neurips-sparse,
  title     = {{Sparse Additive Text Models with Low Rank Background}},
  author    = {Shi, Lei},
  booktitle = {Neural Information Processing Systems},
  year      = {2013},
  pages     = {172-180},
  url       = {https://mlanthology.org/neurips/2013/shi2013neurips-sparse/}
}