On Inference for the Support Vector Machine
Abstract
The linear support vector machine has a parametrised decision boundary. The paper considers inference for the corresponding parameters, which indicate the effects of individual variables on the decision boundary. The proposed inference is via a convolution-smoothed version of the SVM loss function, this having several inferential advantages over the original SVM, whose associated loss function is not everywhere differentiable. Notably, convolution-smoothing comes with non-asymptotic theoretical guarantees, including a distributional approximation to the parameter estimator that scales more favourably with the dimension of the feature vector. The differentiability of the loss function produces other advantages in some settings; for instance, by facilitating the inclusion of penalties or the synthesis of information from a large number of small samples. The paper closes by relating the linear SVM parameters to those of some probability models for binary outcomes.
Cite
Text
Rybak et al. "On Inference for the Support Vector Machine." Journal of Machine Learning Research, 2025.Markdown
[Rybak et al. "On Inference for the Support Vector Machine." Journal of Machine Learning Research, 2025.](https://mlanthology.org/jmlr/2025/rybak2025jmlr-inference/)BibTeX
@article{rybak2025jmlr-inference,
title = {{On Inference for the Support Vector Machine}},
author = {Rybak, Jakub and Battey, Heather and Zhou, Wen-Xin},
journal = {Journal of Machine Learning Research},
year = {2025},
pages = {1-54},
volume = {26},
url = {https://mlanthology.org/jmlr/2025/rybak2025jmlr-inference/}
}