Generalization Bounds for Learning Kernels
Abstract
This paper presents several novel generalization bounds for the problem of learning kernels based on a combinatorial analysis of the Rademacher complexity of the corresponding hypothesis sets. Our bound for learning kernels with a convex combination of p base kernels using L_1 regularization admits only a \sqrt{\log p} dependency on the number of kernels, which is *tight* and considerably more favorable than the previous best bound given for the same problem. We also give a novel bound for learning with a non-negative combination of p base kernels with an L_2 regularization whose dependency on p is also*tight* and only in p^1/4. We present similar results for L_q regularization with other values of q, and outline the relevance of our proof techniques to the analysis of the complexity of the class of linear functions. Experiments with a large number of kernels further validate the behavior of the generalization error as a function of p predicted by our bounds.
Cite
Text
Cortes et al. "Generalization Bounds for Learning Kernels." International Conference on Machine Learning, 2010.Markdown
[Cortes et al. "Generalization Bounds for Learning Kernels." International Conference on Machine Learning, 2010.](https://mlanthology.org/icml/2010/cortes2010icml-generalization/)BibTeX
@inproceedings{cortes2010icml-generalization,
title = {{Generalization Bounds for Learning Kernels}},
author = {Cortes, Corinna and Mohri, Mehryar and Rostamizadeh, Afshin},
booktitle = {International Conference on Machine Learning},
year = {2010},
pages = {247-254},
url = {https://mlanthology.org/icml/2010/cortes2010icml-generalization/}
}