Exponential Convergence Rates in Classification
Abstract
Let ( X , Y ) be a random couple, X being an observable instance and Y ∈ –1,1 being a binary label to be predicted based on an observation of the instance. Let ( X _ i , Y _ i ), i = 1, . . . , n be training data consisting of n independent copies of ( X , Y ). Consider a real valued classifier ${\hat{f}_{n}}$ that minimizes the following penalized empirical risk $\frac{1}{n}\sum\limits_{i=1}^n \ell(Y_{i}f(X_{i})) + \lambda\|f\|^{2} \rightarrow {\rm min}, f\in {\mathcal H}$ over a Hilbert space ${\mathcal H}$ of functions with norm || ·||, ℓ being a convex loss function and λ >0 being a regularization parameter. In particular, ${\mathcal H}$ might be a Sobolev space or a reproducing kernel Hilbert space. We provide some conditions under which the generalization error of the corresponding binary classifier sign $({\hat{f}_{n}})$ converges to the Bayes risk exponentially fast.
Cite
Text
Koltchinskii and Beznosova. "Exponential Convergence Rates in Classification." Annual Conference on Computational Learning Theory, 2005. doi:10.1007/11503415_20Markdown
[Koltchinskii and Beznosova. "Exponential Convergence Rates in Classification." Annual Conference on Computational Learning Theory, 2005.](https://mlanthology.org/colt/2005/koltchinskii2005colt-exponential/) doi:10.1007/11503415_20BibTeX
@inproceedings{koltchinskii2005colt-exponential,
title = {{Exponential Convergence Rates in Classification}},
author = {Koltchinskii, Vladimir and Beznosova, Olexandra},
booktitle = {Annual Conference on Computational Learning Theory},
year = {2005},
pages = {295-307},
doi = {10.1007/11503415_20},
url = {https://mlanthology.org/colt/2005/koltchinskii2005colt-exponential/}
}