PAC-Bayesian Bound for the Conditional Value at Risk

Abstract

Conditional Value at Risk ($\textsc{CVaR}$) is a ``coherent risk measure'' which generalizes expectation (reduced to a boundary parameter setting). Widely used in mathematical finance, it is garnering increasing interest in machine learning as an alternate approach to regularization, and as a means for ensuring fairness. This paper presents a generalization bound for learning algorithms that minimize the $\textsc{CVaR}$ of the empirical loss. The bound is of PAC-Bayesian type and is guaranteed to be small when the empirical $\textsc{CVaR}$ is small. We achieve this by reducing the problem of estimating $\textsc{CVaR}$ to that of merely estimating an expectation. This then enables us, as a by-product, to obtain concentration inequalities for $\textsc{CVaR}$ even when the random variable in question is unbounded.

Cite

Text

Mhammedi et al. "PAC-Bayesian Bound for the Conditional Value at Risk." Neural Information Processing Systems, 2020.

Markdown

[Mhammedi et al. "PAC-Bayesian Bound for the Conditional Value at Risk." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/mhammedi2020neurips-pacbayesian/)

BibTeX

@inproceedings{mhammedi2020neurips-pacbayesian,
  title     = {{PAC-Bayesian Bound for the Conditional Value at Risk}},
  author    = {Mhammedi, Zakaria and Guedj, Benjamin and Williamson, Robert C.},
  booktitle = {Neural Information Processing Systems},
  year      = {2020},
  url       = {https://mlanthology.org/neurips/2020/mhammedi2020neurips-pacbayesian/}
}