Non-Asymptotic Analysis for Nonparametric Testing
Abstract
We develop a non-asymptotic framework for hypothesis testing in nonparametric regression where the true regression function belongs to a Sobolev space. Our statistical guarantees are exact in the sense that Type I and II errors are controlled for any finite sample size. Meanwhile, one proposed test is shown to achieve minimax rate optimality in the asymptotic sense. An important consequence of this non-asymptotic theory is a new and practically useful formula for selecting the optimal smoothing parameter in the testing statistic. Extensions of our results to general reproducing kernel Hilbert spaces and non-Gaussian error regression are also discussed.
Cite
Text
Yang et al. "Non-Asymptotic Analysis for Nonparametric Testing." Conference on Learning Theory, 2020.Markdown
[Yang et al. "Non-Asymptotic Analysis for Nonparametric Testing." Conference on Learning Theory, 2020.](https://mlanthology.org/colt/2020/yang2020colt-nonasymptotic/)BibTeX
@inproceedings{yang2020colt-nonasymptotic,
title = {{Non-Asymptotic Analysis for Nonparametric Testing}},
author = {Yang, Yun and Shang, Zuofeng and Cheng, Guang},
booktitle = {Conference on Learning Theory},
year = {2020},
pages = {3709-3755},
volume = {125},
url = {https://mlanthology.org/colt/2020/yang2020colt-nonasymptotic/}
}