Coordinate Descent Method for Large-Scale L2-Loss Linear Support Vector Machines
Abstract
Linear support vector machines (SVM) are useful for classifying large-scale sparse data. Problems with sparse features are common in applications such as document classification and natural language processing. In this paper, we propose a novel coordinate descent algorithm for training linear SVM with the L2-loss function. At each step, the proposed method minimizes a one-variable sub-problem while fixing other variables. The sub-problem is solved by Newton steps with the line search technique. The procedure globally converges at the linear rate. As each sub-problem involves only values of a corresponding feature, the proposed approach is suitable when accessing a feature is more convenient than accessing an instance. Experiments show that our method is more efficient and stable than state of the art methods such as Pegasos and TRON.
Cite
Text
Chang et al. "Coordinate Descent Method for Large-Scale L2-Loss Linear Support Vector Machines." Journal of Machine Learning Research, 2008.Markdown
[Chang et al. "Coordinate Descent Method for Large-Scale L2-Loss Linear Support Vector Machines." Journal of Machine Learning Research, 2008.](https://mlanthology.org/jmlr/2008/chang2008jmlr-coordinate/)BibTeX
@article{chang2008jmlr-coordinate,
title = {{Coordinate Descent Method for Large-Scale L2-Loss Linear Support Vector Machines}},
author = {Chang, Kai-Wei and Hsieh, Cho-Jui and Lin, Chih-Jen},
journal = {Journal of Machine Learning Research},
year = {2008},
pages = {1369-1398},
volume = {9},
url = {https://mlanthology.org/jmlr/2008/chang2008jmlr-coordinate/}
}