Optimizing F-Measures by Cost-Sensitive Classification

Abstract

We present a theoretical analysis of F-measures for binary, multiclass and multilabel classification. These performance measures are non-linear, but in many scenarios they are pseudo-linear functions of the per-class false negative/false positive rate. Based on this observation, we present a general reduction of F-measure maximization to cost-sensitive classification with unknown costs. We then propose an algorithm with provable guarantees to obtain an approximately optimal classifier for the F-measure by solving a series of cost-sensitive classification problems. The strength of our analysis is to be valid on any dataset and any class of classifiers, extending the existing theoretical results on F-measures, which are asymptotic in nature. We present numerical experiments to illustrate the relative importance of cost asymmetry and thresholding when learning linear classifiers on various F-measure optimization tasks.

Cite

Text

Parambath et al. "Optimizing F-Measures by Cost-Sensitive Classification." Neural Information Processing Systems, 2014.

Markdown

[Parambath et al. "Optimizing F-Measures by Cost-Sensitive Classification." Neural Information Processing Systems, 2014.](https://mlanthology.org/neurips/2014/parambath2014neurips-optimizing/)

BibTeX

@inproceedings{parambath2014neurips-optimizing,
  title     = {{Optimizing F-Measures by Cost-Sensitive Classification}},
  author    = {Parambath, Shameem Puthiya and Usunier, Nicolas and Grandvalet, Yves},
  booktitle = {Neural Information Processing Systems},
  year      = {2014},
  pages     = {2123-2131},
  url       = {https://mlanthology.org/neurips/2014/parambath2014neurips-optimizing/}
}