PAC-Bayesian Bound for the Conditional Value at Risk
Abstract
Conditional Value at Risk ($\textsc{CVaR}$) is a ``coherent risk measure'' which generalizes expectation (reduced to a boundary parameter setting). Widely used in mathematical finance, it is garnering increasing interest in machine learning as an alternate approach to regularization, and as a means for ensuring fairness. This paper presents a generalization bound for learning algorithms that minimize the $\textsc{CVaR}$ of the empirical loss. The bound is of PAC-Bayesian type and is guaranteed to be small when the empirical $\textsc{CVaR}$ is small. We achieve this by reducing the problem of estimating $\textsc{CVaR}$ to that of merely estimating an expectation. This then enables us, as a by-product, to obtain concentration inequalities for $\textsc{CVaR}$ even when the random variable in question is unbounded.
Cite
Text
Mhammedi et al. "PAC-Bayesian Bound for the Conditional Value at Risk." Neural Information Processing Systems, 2020.Markdown
[Mhammedi et al. "PAC-Bayesian Bound for the Conditional Value at Risk." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/mhammedi2020neurips-pacbayesian/)BibTeX
@inproceedings{mhammedi2020neurips-pacbayesian,
title = {{PAC-Bayesian Bound for the Conditional Value at Risk}},
author = {Mhammedi, Zakaria and Guedj, Benjamin and Williamson, Robert C.},
booktitle = {Neural Information Processing Systems},
year = {2020},
url = {https://mlanthology.org/neurips/2020/mhammedi2020neurips-pacbayesian/}
}