$\ell_1$-Regression with Heavy-Tailed Distributions
Abstract
In this paper, we consider the problem of linear regression with heavy-tailed distributions. Different from previous studies that use the squared loss to measure the performance, we choose the absolute loss, which is capable of estimating the conditional median. To address the challenge that both the input and output could be heavy-tailed, we propose a truncated minimization problem, and demonstrate that it enjoys an $O(\sqrt{d/n})$ excess risk, where $d$ is the dimensionality and $n$ is the number of samples. Compared with traditional work on $\ell_1$-regression, the main advantage of our result is that we achieve a high-probability risk bound without exponential moment conditions on the input and output. Furthermore, if the input is bounded, we show that the classical empirical risk minimization is competent for $\ell_1$-regression even when the output is heavy-tailed.
Cite
Text
Zhang and Zhou. "$\ell_1$-Regression with Heavy-Tailed Distributions." Neural Information Processing Systems, 2018.Markdown
[Zhang and Zhou. "$\ell_1$-Regression with Heavy-Tailed Distributions." Neural Information Processing Systems, 2018.](https://mlanthology.org/neurips/2018/zhang2018neurips-1regression/)BibTeX
@inproceedings{zhang2018neurips-1regression,
title = {{$\ell_1$-Regression with Heavy-Tailed Distributions}},
author = {Zhang, Lijun and Zhou, Zhi-Hua},
booktitle = {Neural Information Processing Systems},
year = {2018},
pages = {1076-1086},
url = {https://mlanthology.org/neurips/2018/zhang2018neurips-1regression/}
}