Towards Sharper Generalization Bounds for Structured Prediction

Abstract

In this paper, we investigate the generalization performance of structured prediction learning and obtain state-of-the-art generalization bounds. Our analysis is based on factor graph decomposition of structured prediction algorithms, and we present novel margin guarantees from three different perspectives: Lipschitz continuity, smoothness, and space capacity condition. In the Lipschitz continuity scenario, we improve the square-root dependency on the label set cardinality of existing bounds to a logarithmic dependence. In the smoothness scenario, we provide generalization bounds that are not only a logarithmic dependency on the label set cardinality but a faster convergence rate of order $\mathcal{O}(\frac{1}{n})$ on the sample size $n$. In the space capacity scenario, we obtain bounds that do not depend on the label set cardinality and have faster convergence rates than $\mathcal{O}(\frac{1}{\sqrt{n}})$. In each scenario, applications are provided to suggest that these conditions are easy to be satisfied.

Cite

Text

Li and Liu. "Towards Sharper Generalization Bounds for Structured Prediction." Neural Information Processing Systems, 2021.

Markdown

[Li and Liu. "Towards Sharper Generalization Bounds for Structured Prediction." Neural Information Processing Systems, 2021.](https://mlanthology.org/neurips/2021/li2021neurips-sharper/)

BibTeX

@inproceedings{li2021neurips-sharper,
  title     = {{Towards Sharper Generalization Bounds for Structured Prediction}},
  author    = {Li, Shaojie and Liu, Yong},
  booktitle = {Neural Information Processing Systems},
  year      = {2021},
  url       = {https://mlanthology.org/neurips/2021/li2021neurips-sharper/}
}