Tighter Information-Theoretic Generalization Bounds from Supersamples
Abstract
In this work, we present a variety of novel information-theoretic generalization bounds for learning algorithms, from the supersample setting of Steinke & Zakynthinou (2020)—the setting of the “conditional mutual information” framework. Our development exploits projecting the loss pair (obtained from a training instance and a testing instance) down to a single number and correlating loss values with a Rademacher sequence (and its shifted variants). The presented bounds include square-root bounds, fast-rate bounds, including those based on variance and sharpness, and bounds for interpolating algorithms etc. We show theoretically or empirically that these bounds are tighter than all information-theoretic bounds known to date on the same supersample setting.
Cite
Text
Wang and Mao. "Tighter Information-Theoretic Generalization Bounds from Supersamples." International Conference on Machine Learning, 2023.Markdown
[Wang and Mao. "Tighter Information-Theoretic Generalization Bounds from Supersamples." International Conference on Machine Learning, 2023.](https://mlanthology.org/icml/2023/wang2023icml-tighter/)BibTeX
@inproceedings{wang2023icml-tighter,
title = {{Tighter Information-Theoretic Generalization Bounds from Supersamples}},
author = {Wang, Ziqiao and Mao, Yongyi},
booktitle = {International Conference on Machine Learning},
year = {2023},
pages = {36111-36137},
volume = {202},
url = {https://mlanthology.org/icml/2023/wang2023icml-tighter/}
}