Multiclass Multiple Kernel Learning

Abstract

In many applications it is desirable to learn from several kernels. "Multiple kernel learning" (MKL) allows the practitioner to optimize over linear combinations of kernels. By enforcing sparse coefficients, it also generalizes feature selection to kernel selection. We propose MKL for joint feature maps. This provides a convenient and principled way for MKL with multiclass problems. In addition, we can exploit the joint feature map to learn kernels on output spaces. We show the equivalence of several different primal formulations including different regularizers. We present several optimization methods, and compare a convex quadratically constrained quadratic program (QCQP) and two semi-infinite linear programs (SILPs) on toy data, showing that the SILPs are faster than the QCQP. We then demonstrate the utility of our method by applying the SILP to three real world datasets.

Cite

Text

Zien and Ong. "Multiclass Multiple Kernel Learning." International Conference on Machine Learning, 2007. doi:10.1145/1273496.1273646

Markdown

[Zien and Ong. "Multiclass Multiple Kernel Learning." International Conference on Machine Learning, 2007.](https://mlanthology.org/icml/2007/zien2007icml-multiclass/) doi:10.1145/1273496.1273646

BibTeX

@inproceedings{zien2007icml-multiclass,
  title     = {{Multiclass Multiple Kernel Learning}},
  author    = {Zien, Alexander and Ong, Cheng Soon},
  booktitle = {International Conference on Machine Learning},
  year      = {2007},
  pages     = {1191-1198},
  doi       = {10.1145/1273496.1273646},
  url       = {https://mlanthology.org/icml/2007/zien2007icml-multiclass/}
}