Minimal Kernel Classifiers (Kernel Machines Section)

Abstract

A finite concave minimization algorithm is proposed for constructing kernel classifiers that use a minimal number of data points both in generating and characterizing a classifier. The algorithm is theoretically justified on the basis of linear programming perturbation theory and a leave-one-out error bound as well as effective computational results on seven real world datasets. A nonlinear rectangular kernel is generated by systematically utilizing as few of the data as possible both in training and in characterizing a nonlinear separating surface. This can result in substantial reduction in kernel data-dependence (over 94% in six of the seven public datasets tested on) and with test set correctness equal to that obtained by using a conventional support vector machine classifier that depends on many more data points. This reduction in data dependence results in a much faster classifier that requires less storage. To eliminate data points, the proposed approach makes use of a novel loss function, the "pound" function ()#, which is a linear combination of the 1-norm and the step function that measures both the magnitude and the presence of any error.

Cite

Text

Fung et al. "Minimal Kernel Classifiers     (Kernel Machines Section)." Journal of Machine Learning Research, 2002.

Markdown

[Fung et al. "Minimal Kernel Classifiers     (Kernel Machines Section)." Journal of Machine Learning Research, 2002.](https://mlanthology.org/jmlr/2002/fung2002jmlr-minimal/)

BibTeX

@article{fung2002jmlr-minimal,
  title     = {{Minimal Kernel Classifiers     (Kernel Machines Section)}},
  author    = {Fung, Glenn M. and Mangasarian, Olvi L. and Smola, Alexander J.},
  journal   = {Journal of Machine Learning Research},
  year      = {2002},
  pages     = {303-321},
  volume    = {3},
  url       = {https://mlanthology.org/jmlr/2002/fung2002jmlr-minimal/}
}