On Primal and Dual Sparsity of Markov Networks

Abstract

Sparsity is a desirable property in high dimensional learning. The l<sub>1</sub>-norm regularization can lead to primal sparsity, while max-margin methods achieve dual sparsity. Combining these two methods, an l<sub>1</sub>-norm max-margin Markov network (l<sub>1</sub>-M<sup>3</sup>N) can achieve both types of sparsity. This paper analyzes its connections to the Laplace max-margin Markov network (LapM<sup>3</sup>N), which inherits the dual sparsity of max-margin models but is pseudo-primal sparse, and to a novel adaptive M<sup>3</sup>N (AdapM<sup>3</sup>N). We show that the l<sub>1</sub>-M<sup>3</sup>N is an extreme case of the LapM<sup>3</sup>N, and the l<sub>1</sub>-M<sup>3</sup>N is equivalent to an AdapM<sup>3</sup>N. Based on this equivalence we develop a robust EM-style algorithm for learning an l<sub>1</sub>-M<sup>3</sup>N. We demonstrate the advantages of the simultaneously (pseudo-) primal and dual sparse models over the ones which enjoy either primal or dual sparsity on both synthetic and real data sets.

Cite

Text

Zhu and Xing. "On Primal and Dual Sparsity of Markov Networks." International Conference on Machine Learning, 2009. doi:10.1145/1553374.1553536

Markdown

[Zhu and Xing. "On Primal and Dual Sparsity of Markov Networks." International Conference on Machine Learning, 2009.](https://mlanthology.org/icml/2009/zhu2009icml-primal/) doi:10.1145/1553374.1553536

BibTeX

@inproceedings{zhu2009icml-primal,
  title     = {{On Primal and Dual Sparsity of Markov Networks}},
  author    = {Zhu, Jun and Xing, Eric P.},
  booktitle = {International Conference on Machine Learning},
  year      = {2009},
  pages     = {1265-1272},
  doi       = {10.1145/1553374.1553536},
  url       = {https://mlanthology.org/icml/2009/zhu2009icml-primal/}
}