Direct Convex Relaxations of Sparse SVM

Abstract

Although support vector machines (SVMs) for binary classification give rise to a decision rule that only relies on a subset of the training data points (support vectors), it will in general be based on all available features in the input space. We propose two direct, novel convex relaxations of a nonconvex sparse SVM formulation that explicitly constrains the cardinality of the vector of feature weights. One relaxation results in a quadratically-constrained quadratic program (QCQP), while the second is based on a semidefinite programming (SDP) relaxation. The QCQP formulation can be interpreted as applying an adaptive soft-threshold on the SVM hyperplane, while the SDP formulation learns a weighted inner-product (i.e. a kernel) that results in a sparse hyperplane. Experimental results show an increase in sparsity while conserving the generalization performance compared to a standard as well as a linear programming SVM.

Cite

Text

Chan et al. "Direct Convex Relaxations of Sparse SVM." International Conference on Machine Learning, 2007. doi:10.1145/1273496.1273515

Markdown

[Chan et al. "Direct Convex Relaxations of Sparse SVM." International Conference on Machine Learning, 2007.](https://mlanthology.org/icml/2007/chan2007icml-direct/) doi:10.1145/1273496.1273515

BibTeX

@inproceedings{chan2007icml-direct,
  title     = {{Direct Convex Relaxations of Sparse SVM}},
  author    = {Chan, Antoni B. and Vasconcelos, Nuno and Lanckriet, Gert R. G.},
  booktitle = {International Conference on Machine Learning},
  year      = {2007},
  pages     = {145-153},
  doi       = {10.1145/1273496.1273515},
  url       = {https://mlanthology.org/icml/2007/chan2007icml-direct/}
}