The Last-Step Minimax Algorithm
Abstract
We consider on-line density estimation with a parameterized density from an exponential family. In each trial t the learner predicts a parameter θ _ t . Then it receives an instance x_t chosen by the adversary and incurs loss -ln p ( χ _ t | θ _ t which is the negative log-likelihood of χ _ t w.r.t. the predicted density of the learner. The performance of the learner is measured by the regret defined as the total loss of the learner minus the total loss of the best parameter chosen off-line. We develop an algorithm called the Last-step Minimax Algorithm that predicts with the minimax optimal parameter assuming that the current trial is the last one. For one-dimensional exponential families, we give an explicit form of the prediction of the Last-step Minimax Algorithm and show that its regret is O (ln T ), where T is the number of trials. In particular, for Bernoulli density estimation the Last-step Minimax Algorithm is slightly better than the standard Krichevsky-Trofimov probability estimator
Cite
Text
Takimoto and Warmuth. "The Last-Step Minimax Algorithm." International Conference on Algorithmic Learning Theory, 2000. doi:10.1007/3-540-40992-0_21Markdown
[Takimoto and Warmuth. "The Last-Step Minimax Algorithm." International Conference on Algorithmic Learning Theory, 2000.](https://mlanthology.org/alt/2000/takimoto2000alt-laststep/) doi:10.1007/3-540-40992-0_21BibTeX
@inproceedings{takimoto2000alt-laststep,
title = {{The Last-Step Minimax Algorithm}},
author = {Takimoto, Eiji and Warmuth, Manfred K.},
booktitle = {International Conference on Algorithmic Learning Theory},
year = {2000},
pages = {279-290},
doi = {10.1007/3-540-40992-0_21},
url = {https://mlanthology.org/alt/2000/takimoto2000alt-laststep/}
}