Generalisation Error Bounds for Sparse Linear Classifiers
Abstract
We provide small sample size bounds on the generalisation error of linear classifiers that are sparse in their dual representation given by the expansion coefficients of the weight vector in terms of the training data. These results theoretically justify algorithms like the Support Vector Machine, the Relevance Vector Machine and K-nearest-neighbour. The bounds are a-posteriori bounds to be evaluated after learning when the attained level of sparsity is known. In a PAC-Bayesian style prior knowledge about the expected sparsity is incorporated into the bounds. The proofs avoid the use of double sample arguments by taking into account the sparsity that leaves unused training points for the evaluation of classifiers. We furthermore give a PAC-Bayesian bound on the average generalisation error over subsets of parameter space that may pave the way combining sparsity in the expansion coefficients and margin in a single bound. Finally, reinterpreting a mistake bound for the classical perceptr...
Cite
Text
Graepel et al. "Generalisation Error Bounds for Sparse Linear Classifiers." Annual Conference on Computational Learning Theory, 2000.Markdown
[Graepel et al. "Generalisation Error Bounds for Sparse Linear Classifiers." Annual Conference on Computational Learning Theory, 2000.](https://mlanthology.org/colt/2000/graepel2000colt-generalisation/)BibTeX
@inproceedings{graepel2000colt-generalisation,
title = {{Generalisation Error Bounds for Sparse Linear Classifiers}},
author = {Graepel, Thore and Herbrich, Ralf and Shawe-Taylor, John},
booktitle = {Annual Conference on Computational Learning Theory},
year = {2000},
pages = {298-303},
url = {https://mlanthology.org/colt/2000/graepel2000colt-generalisation/}
}