Margin and Radius Based Multiple Kernel Learning

Abstract

A serious drawback of kernel methods, and Support Vector Machines (SVM) in particular, is the difficulty in choosing a suitable kernel function for a given dataset. One of the approaches proposed to address this problem is Multiple Kernel Learning (MKL) in which several kernels are combined adaptively for a given dataset. Many of the existing MKL methods use the SVM objective function and try to find a linear combination of basic kernels such that the separating margin between the classes is maximized. However, these methods ignore the fact that the theoretical error bound depends not only on the margin, but also on the radius of the smallest sphere that contains all the training instances. We present a novel MKL algorithm that optimizes the error bound taking account of both the margin and the radius. The empirical results show that the proposed method compares favorably with other state-of-the-art MKL methods.

Cite

Text

Do et al. "Margin and Radius Based Multiple Kernel Learning." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2009. doi:10.1007/978-3-642-04180-8_39

Markdown

[Do et al. "Margin and Radius Based Multiple Kernel Learning." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2009.](https://mlanthology.org/ecmlpkdd/2009/do2009ecmlpkdd-margin/) doi:10.1007/978-3-642-04180-8_39

BibTeX

@inproceedings{do2009ecmlpkdd-margin,
  title     = {{Margin and Radius Based Multiple Kernel Learning}},
  author    = {Do, Huyen and Kalousis, Alexandros and Woznica, Adam and Hilario, Melanie},
  booktitle = {European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases},
  year      = {2009},
  pages     = {330-343},
  doi       = {10.1007/978-3-642-04180-8_39},
  url       = {https://mlanthology.org/ecmlpkdd/2009/do2009ecmlpkdd-margin/}
}