Fine-Grained Generalization Analysis of Vector-Valued Learning

Abstract

Many fundamental machine learning tasks can be formulated as a problem of learning with vector-valued functions, where we learn multiple scalar-valued functions together. Although there is some generalization analysis on different specific algorithms under the empirical risk minimization principle, a unifying analysis of vector-valued learning under a regularization framework is still lacking. In this paper, we initiate the generalization analysis of regularized vector-valued learning algorithms by presenting bounds with a mild dependency on the output dimension and a fast rate on the sample size. Our discussions relax the existing assumptions on the restrictive constraint of hypothesis spaces, smoothness of loss functions and low-noise condition. To understand the interaction between optimization and learning, we further use our results to derive the first generalization bounds for stochastic gradient descent with vector-valued functions. We apply our general results to multi-class classification and multi-label classification, which yield the first bounds with a logarithmic dependency on the output dimension for extreme multi-label classification with the Frobenius regularization. As a byproduct, we derive a Rademacher complexity bound for loss function classes defined in terms of a general strongly convex function.

Cite

Text

Wu et al. "Fine-Grained Generalization Analysis of Vector-Valued Learning." AAAI Conference on Artificial Intelligence, 2021. doi:10.1609/AAAI.V35I12.17238

Markdown

[Wu et al. "Fine-Grained Generalization Analysis of Vector-Valued Learning." AAAI Conference on Artificial Intelligence, 2021.](https://mlanthology.org/aaai/2021/wu2021aaai-fine/) doi:10.1609/AAAI.V35I12.17238

BibTeX

@inproceedings{wu2021aaai-fine,
  title     = {{Fine-Grained Generalization Analysis of Vector-Valued Learning}},
  author    = {Wu, Liang and Ledent, Antoine and Lei, Yunwen and Kloft, Marius},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2021},
  pages     = {10338-10346},
  doi       = {10.1609/AAAI.V35I12.17238},
  url       = {https://mlanthology.org/aaai/2021/wu2021aaai-fine/}
}