Optimal Rates of Kernel Ridge Regression Under Source Condition in Large Dimensions
Abstract
Motivated by studies of neural networks, particularly the neural tangent kernel theory, we investigate the large-dimensional behavior of kernel ridge regression, where the sample size satisfies $n $ is proportion to $ d^{\gamma}$ for some $\gamma > 0$. Given a reproducing kernel Hilbert space $H$ associated with an inner product kernel defined on the unit sphere $S^{d}$, we assume that the true function $f_{\rho}^{*}$ belongs to the interpolation space $[H]^{s}$ for some $s>0$ (source condition). We first establish the exact order (both upper and lower bounds) of the generalization error of KRR for the optimally chosen regularization parameter $\lambda$. Furthermore, we show that KRR is minimax optimal when $01$, KRR fails to achieve minimax optimality, exhibiting the saturation effect. Our results illustrate that the convergence rate with respect to dimension $d$ varying along $\gamma$ exhibits a periodic plateau behavior, and the convergence rate with respect to sample size $n$ exhibits a multiple descent behavior. Interestingly, our work unifies several recent studies on kernel regression in the large-dimensional setting, which correspond to $s=0$ and $s=1$, respectively.
Cite
Text
Zhang et al. "Optimal Rates of Kernel Ridge Regression Under Source Condition in Large Dimensions." Journal of Machine Learning Research, 2025.Markdown
[Zhang et al. "Optimal Rates of Kernel Ridge Regression Under Source Condition in Large Dimensions." Journal of Machine Learning Research, 2025.](https://mlanthology.org/jmlr/2025/zhang2025jmlr-optimal/)BibTeX
@article{zhang2025jmlr-optimal,
title = {{Optimal Rates of Kernel Ridge Regression Under Source Condition in Large Dimensions}},
author = {Zhang, Haobo and Li, Yicheng and Lu, Weihao and Lin, Qian},
journal = {Journal of Machine Learning Research},
year = {2025},
pages = {1-63},
volume = {26},
url = {https://mlanthology.org/jmlr/2025/zhang2025jmlr-optimal/}
}