Simplified PAC-Bayesian Margin Bounds
Abstract
The theoretical understanding of support vector machines is largely based on margin bounds for linear classifiers with unit-norm weight vectors and unit-norm feature vectors. Unit-norm margin bounds have been proved previously using fat-shattering arguments and Rademacher complexity. Recently Langford and Shawe-Taylor proved a dimension-independent unit-norm margin bound using a relatively simple PAC-Bayesian argument. Unfortunately, the Langford-Shawe-Taylor bound is stated in a variational form making direct comparison to fat-shattering bounds difficult. This paper provides an explicit solution to the variational problem implicit in the Langford-Shawe-Taylor bound and shows that the PAC-Bayesian margin bounds are significantly tighter. Because a PAC-Bayesian bound is derived from a particular prior distribution over hypotheses, a PAC-Bayesian margin bound also seems to provide insight into the nature of the learning bias underlying the bound.
Cite
Text
McAllester. "Simplified PAC-Bayesian Margin Bounds." Annual Conference on Computational Learning Theory, 2003. doi:10.1007/978-3-540-45167-9_16Markdown
[McAllester. "Simplified PAC-Bayesian Margin Bounds." Annual Conference on Computational Learning Theory, 2003.](https://mlanthology.org/colt/2003/mcallester2003colt-simplified/) doi:10.1007/978-3-540-45167-9_16BibTeX
@inproceedings{mcallester2003colt-simplified,
title = {{Simplified PAC-Bayesian Margin Bounds}},
author = {McAllester, David A.},
booktitle = {Annual Conference on Computational Learning Theory},
year = {2003},
pages = {203-215},
doi = {10.1007/978-3-540-45167-9_16},
url = {https://mlanthology.org/colt/2003/mcallester2003colt-simplified/}
}