Large Scale Kernel Regression via Linear Programming

Abstract

The problem of tolerant data fitting by a nonlinear surface, induced by a kernel-based support vector machine is formulated as a linear program with fewer number of variables than that of other linear programming formulations. A generalization of the linear programming chunking algorithm for arbitrary kernels is implemented for solving problems with very large datasets wherein chunking is performed on both data points and problem variables. The proposed approach tolerates a small error, which is adjusted parametrically, while fitting the given data. This leads to improved fitting of noisy data (over ordinary least error solutions) as demonstrated computationally. Comparative numerical results indicate an average time reduction as high as 26.0% over other formulations, with a maximal time reduction of 79.7%. Additionally, linear programs with as many as 16,000 data points and more than a billion nonzero matrix elements are solved.

Cite

Text

Mangasarian and Musicant. "Large Scale Kernel Regression via Linear Programming." Machine Learning, 2002. doi:10.1023/A:1012422931930

Markdown

[Mangasarian and Musicant. "Large Scale Kernel Regression via Linear Programming." Machine Learning, 2002.](https://mlanthology.org/mlj/2002/mangasarian2002mlj-large/) doi:10.1023/A:1012422931930

BibTeX

@article{mangasarian2002mlj-large,
  title     = {{Large Scale Kernel Regression via Linear Programming}},
  author    = {Mangasarian, Olvi L. and Musicant, David R.},
  journal   = {Machine Learning},
  year      = {2002},
  pages     = {255-269},
  doi       = {10.1023/A:1012422931930},
  volume    = {46},
  url       = {https://mlanthology.org/mlj/2002/mangasarian2002mlj-large/}
}