Direct Convex Relaxations of Sparse SVM
Abstract
Although support vector machines (SVMs) for binary classification give rise to a decision rule that only relies on a subset of the training data points (support vectors), it will in general be based on all available features in the input space. We propose two direct, novel convex relaxations of a nonconvex sparse SVM formulation that explicitly constrains the cardinality of the vector of feature weights. One relaxation results in a quadratically-constrained quadratic program (QCQP), while the second is based on a semidefinite programming (SDP) relaxation. The QCQP formulation can be interpreted as applying an adaptive soft-threshold on the SVM hyperplane, while the SDP formulation learns a weighted inner-product (i.e. a kernel) that results in a sparse hyperplane. Experimental results show an increase in sparsity while conserving the generalization performance compared to a standard as well as a linear programming SVM.
Cite
Text
Chan et al. "Direct Convex Relaxations of Sparse SVM." International Conference on Machine Learning, 2007. doi:10.1145/1273496.1273515Markdown
[Chan et al. "Direct Convex Relaxations of Sparse SVM." International Conference on Machine Learning, 2007.](https://mlanthology.org/icml/2007/chan2007icml-direct/) doi:10.1145/1273496.1273515BibTeX
@inproceedings{chan2007icml-direct,
title = {{Direct Convex Relaxations of Sparse SVM}},
author = {Chan, Antoni B. and Vasconcelos, Nuno and Lanckriet, Gert R. G.},
booktitle = {International Conference on Machine Learning},
year = {2007},
pages = {145-153},
doi = {10.1145/1273496.1273515},
url = {https://mlanthology.org/icml/2007/chan2007icml-direct/}
}