Confidence Intervals and Hypothesis Testing for High-Dimensional Statistical Models
Abstract
Fitting high-dimensional statistical models often requires the use of non-linear parameter estimation procedures. As a consequence, it is generally impossible to obtain an exact characterization of the probability distribution of the parameter estimates. This in turn implies that it is extremely challenging to quantify the `uncertainty' associated with a certain parameter estimate. Concretely, no commonly accepted procedure exists for computing classical measures of uncertainty and statistical significance as confidence intervals or p-values. We consider here a broad class of regression problems, and propose an efficient algorithm for constructing confidence intervals and p-values. The resulting confidence intervals have nearly optimal size. When testing for the null hypothesis that a certain parameter is vanishing, our method has nearly optimal power. Our approach is based on constructing a `de-biased' version of regularized M-estimators. The new construction improves over recent work in the field in that it does not assume a special structure on the design matrix. Furthermore, proofs are remarkably simple. We test our method on a diabetes prediction problem.
Cite
Text
Javanmard and Montanari. "Confidence Intervals and Hypothesis Testing for High-Dimensional Statistical Models." Neural Information Processing Systems, 2013.Markdown
[Javanmard and Montanari. "Confidence Intervals and Hypothesis Testing for High-Dimensional Statistical Models." Neural Information Processing Systems, 2013.](https://mlanthology.org/neurips/2013/javanmard2013neurips-confidence/)BibTeX
@inproceedings{javanmard2013neurips-confidence,
title = {{Confidence Intervals and Hypothesis Testing for High-Dimensional Statistical Models}},
author = {Javanmard, Adel and Montanari, Andrea},
booktitle = {Neural Information Processing Systems},
year = {2013},
pages = {1187-1195},
url = {https://mlanthology.org/neurips/2013/javanmard2013neurips-confidence/}
}