Learning Bounds for Support Vector Machines with Learned Kernels
Abstract
Consider the problem of learning a kernel for use in SVM classification. We bound the estimation error of a large margin classifier when the kernel, relative to which this margin is defined, is chosen from a family of kernels based on the training sample. For a kernel family with pseudodimension d _ φ , we present a bound of $\sqrt{\tilde{\mathcal{O}}{({d_{\phi}}+1/\gamma^2)}/n}$ on the estimation error for SVMs with margin γ . This is the first bound in which the relation between the margin term and the family-of-kernels term is additive rather then multiplicative. The pseudodimension of families of linear combinations of base kernels is the number of base kernels. Unlike in previous (multiplicative) bounds, there is no non-negativity requirement on the coefficients of the linear combinations. We also give simple bounds on the pseudodimension for families of Gaussian kernels.
Cite
Text
Srebro and Ben-David. "Learning Bounds for Support Vector Machines with Learned Kernels." Annual Conference on Computational Learning Theory, 2006. doi:10.1007/11776420_15Markdown
[Srebro and Ben-David. "Learning Bounds for Support Vector Machines with Learned Kernels." Annual Conference on Computational Learning Theory, 2006.](https://mlanthology.org/colt/2006/srebro2006colt-learning/) doi:10.1007/11776420_15BibTeX
@inproceedings{srebro2006colt-learning,
title = {{Learning Bounds for Support Vector Machines with Learned Kernels}},
author = {Srebro, Nathan and Ben-David, Shai},
booktitle = {Annual Conference on Computational Learning Theory},
year = {2006},
pages = {169-183},
doi = {10.1007/11776420_15},
url = {https://mlanthology.org/colt/2006/srebro2006colt-learning/}
}