Optimistic Bounds for Multi-Output Learning

Abstract

We investigate the challenge of multi-output learning, where the goal is to learn a vector-valued function based on a supervised data set. This includes a range of important problems in Machine Learning including multi-target regression, multi-class classification and multi-label classification. We begin our analysis by introducing the self-bounding Lipschitz condition for multi-output loss functions, which interpolates continuously between a classical Lipschitz condition and a multi-dimensional analogue of a smoothness condition. We then show that the self-bounding Lipschitz condition gives rise to optimistic bounds for multi-output learning, which attain the minimax optimal rate up to logarithmic factors. The proof exploits local Rademacher complexity combined with a powerful minoration inequality due to Srebro, Sridharan and Tewari. As an application we derive a state-of-the-art generalisation bound for multi-class gradient boosting.

Cite

Text

Reeve and Kaban. "Optimistic Bounds for Multi-Output Learning." International Conference on Machine Learning, 2020.

Markdown

[Reeve and Kaban. "Optimistic Bounds for Multi-Output Learning." International Conference on Machine Learning, 2020.](https://mlanthology.org/icml/2020/reeve2020icml-optimistic/)

BibTeX

@inproceedings{reeve2020icml-optimistic,
  title     = {{Optimistic Bounds for Multi-Output Learning}},
  author    = {Reeve, Henry and Kaban, Ata},
  booktitle = {International Conference on Machine Learning},
  year      = {2020},
  pages     = {8030-8040},
  volume    = {119},
  url       = {https://mlanthology.org/icml/2020/reeve2020icml-optimistic/}
}