Fine-Grained Generalization Analysis of Structured Output Prediction
Abstract
In machine learning we often encounter structured output prediction problems (SOPPs), i.e. problems where the output space admits a rich internal structure. Application domains where SOPPs naturally occur include natural language processing, speech recognition, and computer vision. Typical SOPPs have an extremely large label set, which grows exponentially as a function of the size of the output. Existing generalization analysis implies generalization bounds with at least a square-root dependency on the cardinality d of the label set, which can be vacuous in practice. In this paper, we significantly improve the state of the art by developing novel high-probability bounds with a logarithmic dependency on d. Furthermore, we leverage the lens of algorithmic stability to develop generalization bounds in expectation without any dependency on d. Our results therefore build a solid theoretical foundation for learning in large-scale SOPPs. Furthermore, we extend our results to learning with weakly dependent data.
Cite
Text
Mustafa et al. "Fine-Grained Generalization Analysis of Structured Output Prediction." International Joint Conference on Artificial Intelligence, 2021. doi:10.24963/IJCAI.2021/391Markdown
[Mustafa et al. "Fine-Grained Generalization Analysis of Structured Output Prediction." International Joint Conference on Artificial Intelligence, 2021.](https://mlanthology.org/ijcai/2021/mustafa2021ijcai-fine/) doi:10.24963/IJCAI.2021/391BibTeX
@inproceedings{mustafa2021ijcai-fine,
title = {{Fine-Grained Generalization Analysis of Structured Output Prediction}},
author = {Mustafa, Waleed and Lei, Yunwen and Ledent, Antoine and Kloft, Marius},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2021},
pages = {2841-2847},
doi = {10.24963/IJCAI.2021/391},
url = {https://mlanthology.org/ijcai/2021/mustafa2021ijcai-fine/}
}