A Minimax Approach to Supervised Learning
Abstract
Given a task of predicting Y from X, a loss function L, and a set of probability distributions Gamma on (X,Y), what is the optimal decision rule minimizing the worst-case expected loss over Gamma? In this paper, we address this question by introducing a generalization of the maximum entropy principle. Applying this principle to sets of distributions with marginal on X constrained to be the empirical marginal, we provide a minimax interpretation of the maximum likelihood problem over generalized linear models as well as some popular regularization schemes. For quadratic and logarithmic loss functions we revisit well-known linear and logistic regression models. Moreover, for the 0-1 loss we derive a classifier which we call the minimax SVM. The minimax SVM minimizes the worst-case expected 0-1 loss over the proposed Gamma by solving a tractable optimization problem. We perform several numerical experiments to show the power of the minimax SVM in outperforming the SVM.
Cite
Text
Farnia and Tse. "A Minimax Approach to Supervised Learning." Neural Information Processing Systems, 2016.Markdown
[Farnia and Tse. "A Minimax Approach to Supervised Learning." Neural Information Processing Systems, 2016.](https://mlanthology.org/neurips/2016/farnia2016neurips-minimax/)BibTeX
@inproceedings{farnia2016neurips-minimax,
title = {{A Minimax Approach to Supervised Learning}},
author = {Farnia, Farzan and Tse, David},
booktitle = {Neural Information Processing Systems},
year = {2016},
pages = {4240-4248},
url = {https://mlanthology.org/neurips/2016/farnia2016neurips-minimax/}
}