On Primal and Dual Sparsity of Markov Networks
Abstract
Sparsity is a desirable property in high dimensional learning. The l<sub>1</sub>-norm regularization can lead to primal sparsity, while max-margin methods achieve dual sparsity. Combining these two methods, an l<sub>1</sub>-norm max-margin Markov network (l<sub>1</sub>-M<sup>3</sup>N) can achieve both types of sparsity. This paper analyzes its connections to the Laplace max-margin Markov network (LapM<sup>3</sup>N), which inherits the dual sparsity of max-margin models but is pseudo-primal sparse, and to a novel adaptive M<sup>3</sup>N (AdapM<sup>3</sup>N). We show that the l<sub>1</sub>-M<sup>3</sup>N is an extreme case of the LapM<sup>3</sup>N, and the l<sub>1</sub>-M<sup>3</sup>N is equivalent to an AdapM<sup>3</sup>N. Based on this equivalence we develop a robust EM-style algorithm for learning an l<sub>1</sub>-M<sup>3</sup>N. We demonstrate the advantages of the simultaneously (pseudo-) primal and dual sparse models over the ones which enjoy either primal or dual sparsity on both synthetic and real data sets.
Cite
Text
Zhu and Xing. "On Primal and Dual Sparsity of Markov Networks." International Conference on Machine Learning, 2009. doi:10.1145/1553374.1553536Markdown
[Zhu and Xing. "On Primal and Dual Sparsity of Markov Networks." International Conference on Machine Learning, 2009.](https://mlanthology.org/icml/2009/zhu2009icml-primal/) doi:10.1145/1553374.1553536BibTeX
@inproceedings{zhu2009icml-primal,
title = {{On Primal and Dual Sparsity of Markov Networks}},
author = {Zhu, Jun and Xing, Eric P.},
booktitle = {International Conference on Machine Learning},
year = {2009},
pages = {1265-1272},
doi = {10.1145/1553374.1553536},
url = {https://mlanthology.org/icml/2009/zhu2009icml-primal/}
}