Mathematical Characterization of Better-than-Random Multiclass Models
Abstract
A binary supervised model outperforms chance if and only if the determinant of the confusion matrix is positive. This is equivalent to saying that the associated point in the ROC space is above the random guessing line. This also means that Youden's J, Cohen's $\kappa$ and Matthews' correlation coefficient are positive. We extend these results to any number of classes: for a target variable with $m \geq 2$ classes, we show that a model does better than chance if and only if the entries of the confusion matrix verify $m(m-1)$ homogeneous polynomial inequalities of degree 2, which can be expressed using generalized likelihood ratios. We also obtain a more theoretical formulation: a model does better than chance if and only if it is a maximum likelihood estimator of the target variable. When this is the case, we find that the multiclass versions of the previous metrics remain positive. If $m>2$, we notice that no-skill classifiers are only a small part of the topological boundary between better-than-random models and bad models. For $m=3$, we show that bad models occupy exactly 90\% of the ROC space, far more than the 50\% of the two-class problems. Finally, we propose to define weak multiclass classifiers by conditions on these generalized likelihood ratios.
Cite
Text
Foulle. "Mathematical Characterization of Better-than-Random Multiclass Models." Transactions on Machine Learning Research, 2025.Markdown
[Foulle. "Mathematical Characterization of Better-than-Random Multiclass Models." Transactions on Machine Learning Research, 2025.](https://mlanthology.org/tmlr/2025/foulle2025tmlr-mathematical/)BibTeX
@article{foulle2025tmlr-mathematical,
title = {{Mathematical Characterization of Better-than-Random Multiclass Models}},
author = {Foulle, Sébastien},
journal = {Transactions on Machine Learning Research},
year = {2025},
url = {https://mlanthology.org/tmlr/2025/foulle2025tmlr-mathematical/}
}