Multi-Label Learning with Stronger Consistency Guarantees
Abstract
We present a detailed study of surrogate losses and algorithms for multi-label learning, supported by $H$-consistency bounds. We first show that, for the simplest form of multi-label loss (the popular Hamming loss), the well-known consistent binary relevance surrogate suffers from a sub-optimal dependency on the number of labels in terms of $H$-consistency bounds, when using smooth losses such as logistic losses. Furthermore, this loss function fails to account for label correlations. To address these drawbacks, we introduce a novel surrogate loss, *multi-label logistic loss*, that accounts for label correlations and benefits from label-independent $H$-consistency bounds. We then broaden our analysis to cover a more extensive family of multi-label losses, including all common ones and a new extension defined based on linear-fractional functions with respect to the confusion matrix. We also extend our multi-label logistic losses to more comprehensive multi-label comp-sum losses, adapting comp-sum losses from standard classification to the multi-label learning. We prove that this family of surrogate losses benefits from $H$-consistency bounds, and thus Bayes-consistency, across any general multi-label loss. Our work thus proposes a unified surrogate loss framework benefiting from strong consistency guarantees for any multi-label loss, significantly expanding upon previous work which only established Bayes-consistency and for specific loss functions. Additionally, we adapt constrained losses from standard classification to multi-label constrained losses in a similar way, which also benefit from $H$-consistency bounds and thus Bayes-consistency for any multi-label loss. We further describe efficient gradient computation algorithms for minimizing the multi-label logistic loss.
Cite
Text
Mao et al. "Multi-Label Learning with Stronger Consistency Guarantees." Neural Information Processing Systems, 2024. doi:10.52202/079017-0078Markdown
[Mao et al. "Multi-Label Learning with Stronger Consistency Guarantees." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/mao2024neurips-multilabel/) doi:10.52202/079017-0078BibTeX
@inproceedings{mao2024neurips-multilabel,
title = {{Multi-Label Learning with Stronger Consistency Guarantees}},
author = {Mao, Anqi and Mohri, Mehryar and Zhong, Yutao},
booktitle = {Neural Information Processing Systems},
year = {2024},
doi = {10.52202/079017-0078},
url = {https://mlanthology.org/neurips/2024/mao2024neurips-multilabel/}
}