Optimizing F-Measures by Cost-Sensitive Classification
Abstract
We present a theoretical analysis of F-measures for binary, multiclass and multilabel classification. These performance measures are non-linear, but in many scenarios they are pseudo-linear functions of the per-class false negative/false positive rate. Based on this observation, we present a general reduction of F-measure maximization to cost-sensitive classification with unknown costs. We then propose an algorithm with provable guarantees to obtain an approximately optimal classifier for the F-measure by solving a series of cost-sensitive classification problems. The strength of our analysis is to be valid on any dataset and any class of classifiers, extending the existing theoretical results on F-measures, which are asymptotic in nature. We present numerical experiments to illustrate the relative importance of cost asymmetry and thresholding when learning linear classifiers on various F-measure optimization tasks.
Cite
Text
Parambath et al. "Optimizing F-Measures by Cost-Sensitive Classification." Neural Information Processing Systems, 2014.Markdown
[Parambath et al. "Optimizing F-Measures by Cost-Sensitive Classification." Neural Information Processing Systems, 2014.](https://mlanthology.org/neurips/2014/parambath2014neurips-optimizing/)BibTeX
@inproceedings{parambath2014neurips-optimizing,
title = {{Optimizing F-Measures by Cost-Sensitive Classification}},
author = {Parambath, Shameem Puthiya and Usunier, Nicolas and Grandvalet, Yves},
booktitle = {Neural Information Processing Systems},
year = {2014},
pages = {2123-2131},
url = {https://mlanthology.org/neurips/2014/parambath2014neurips-optimizing/}
}