Simplified PAC-Bayesian Margin Bounds

Abstract

The theoretical understanding of support vector machines is largely based on margin bounds for linear classifiers with unit-norm weight vectors and unit-norm feature vectors. Unit-norm margin bounds have been proved previously using fat-shattering arguments and Rademacher complexity. Recently Langford and Shawe-Taylor proved a dimension-independent unit-norm margin bound using a relatively simple PAC-Bayesian argument. Unfortunately, the Langford-Shawe-Taylor bound is stated in a variational form making direct comparison to fat-shattering bounds difficult. This paper provides an explicit solution to the variational problem implicit in the Langford-Shawe-Taylor bound and shows that the PAC-Bayesian margin bounds are significantly tighter. Because a PAC-Bayesian bound is derived from a particular prior distribution over hypotheses, a PAC-Bayesian margin bound also seems to provide insight into the nature of the learning bias underlying the bound.

Cite

Text

McAllester. "Simplified PAC-Bayesian Margin Bounds." Annual Conference on Computational Learning Theory, 2003. doi:10.1007/978-3-540-45167-9_16

Markdown

[McAllester. "Simplified PAC-Bayesian Margin Bounds." Annual Conference on Computational Learning Theory, 2003.](https://mlanthology.org/colt/2003/mcallester2003colt-simplified/) doi:10.1007/978-3-540-45167-9_16

BibTeX

@inproceedings{mcallester2003colt-simplified,
  title     = {{Simplified PAC-Bayesian Margin Bounds}},
  author    = {McAllester, David A.},
  booktitle = {Annual Conference on Computational Learning Theory},
  year      = {2003},
  pages     = {203-215},
  doi       = {10.1007/978-3-540-45167-9_16},
  url       = {https://mlanthology.org/colt/2003/mcallester2003colt-simplified/}
}