Information Processing Equalities and the Information–Risk Bridge

Abstract

We introduce two new classes of measures of information for statistical experiments which generalise and subsume φ-divergences, integral probability metrics, N-distances (MMD), and (f,Γ) divergences between two or more distributions. This enables us to derive a simple geometrical relationship between measures of information and the Bayes risk of a statistical decision problem, thus extending the variational φ-divergence representation to multiple distributions in an entirely symmetric manner. The new families of divergence are closed under the action of Markov operators which yields an information processing equality which is a refinement and generalisation of the classical information processing inequality. This equality gives insight into the significance of the choice of the hypothesis class in classical risk minimization.

Cite

Text

Williamson and Cranko. "Information Processing Equalities and the Information–Risk Bridge." Journal of Machine Learning Research, 2024.

Markdown

[Williamson and Cranko. "Information Processing Equalities and the Information–Risk Bridge." Journal of Machine Learning Research, 2024.](https://mlanthology.org/jmlr/2024/williamson2024jmlr-information/)

BibTeX

@article{williamson2024jmlr-information,
  title     = {{Information Processing Equalities and the Information–Risk Bridge}},
  author    = {Williamson, Robert C. and Cranko, Zac},
  journal   = {Journal of Machine Learning Research},
  year      = {2024},
  pages     = {1-53},
  volume    = {25},
  url       = {https://mlanthology.org/jmlr/2024/williamson2024jmlr-information/}
}