Cost-Sensitive Classification with Cost Uncertainty: Do We Need Surrogate Losses?

Abstract

Abstract In many binary classification applications, the costs of false positives and negatives are imbalanced. Furthermore, there is often uncertainty about the exact costs of these errors. A natural measure-of-interest to be minimised in such scenarios is the expected misclassification cost. We identify many situations where this measure has analytic gradients, and thus it can be used as a training loss and optimised directly using empirical risk minimisation. In particular, we derive such losses from the Beta, Gamma and Gaussian distributions to model different kinds of cost uncertainty. The Beta family includes commonly used losses such as cross-entropy, squared error and 0–1 loss as special cases. The question then arises as to when it is appropriate to directly optimize the measure-of-interest, versus using a standard surrogate like cross-entropy or focal loss during training. After revisiting the theory of surrogate losses, proper losses and cost-sensitive learning to obtain good candidate surrogates out of derived families, we conduct an empirical comparison of derived training losses that, to our knowledge, were never tried on deep neural networks before, with the aim to minimise cost-sensitive measures-of-interest. The findings show that using Beta losses in training leads to improved performance compared to traditional training objectives like cross-entropy, label smoothing, and focal loss. This improvement is seen not only in terms of misclassification cost metrics, but (perhaps surprisingly) also in conventional metrics such as accuracy, mean squared error, and the area under the ROC curve.

Cite

Text

Komisarenko and Kull. "Cost-Sensitive Classification with Cost Uncertainty: Do We Need Surrogate Losses?." Machine Learning, 2025. doi:10.1007/S10994-024-06634-8

Markdown

[Komisarenko and Kull. "Cost-Sensitive Classification with Cost Uncertainty: Do We Need Surrogate Losses?." Machine Learning, 2025.](https://mlanthology.org/mlj/2025/komisarenko2025mlj-costsensitive/) doi:10.1007/S10994-024-06634-8

BibTeX

@article{komisarenko2025mlj-costsensitive,
  title     = {{Cost-Sensitive Classification with Cost Uncertainty: Do We Need Surrogate Losses?}},
  author    = {Komisarenko, Viacheslav and Kull, Meelis},
  journal   = {Machine Learning},
  year      = {2025},
  pages     = {132},
  doi       = {10.1007/S10994-024-06634-8},
  volume    = {114},
  url       = {https://mlanthology.org/mlj/2025/komisarenko2025mlj-costsensitive/}
}