Efficient Learning Equilibrium
Abstract
We introduce efficient learning equilibrium (ELE), a normative ap(cid:173) proach to learning in non cooperative settings. In ELE, the learn(cid:173) ing algorithms themselves are required to be in equilibrium. In addition, the learning algorithms arrive at a desired value after polynomial time, and deviations from a prescribed ELE become ir(cid:173) rational after polynomial time. We prove the existence of an ELE in the perfect monitoring setting, where the desired value is the expected payoff in a Nash equilibrium. We also show that an ELE does not always exist in the imperfect monitoring case. Yet, it exists in the special case of common-interest games. Finally, we extend our results to general stochastic games.
Cite
Text
Brafman and Tennenholtz. "Efficient Learning Equilibrium." Neural Information Processing Systems, 2002.Markdown
[Brafman and Tennenholtz. "Efficient Learning Equilibrium." Neural Information Processing Systems, 2002.](https://mlanthology.org/neurips/2002/brafman2002neurips-efficient/)BibTeX
@inproceedings{brafman2002neurips-efficient,
title = {{Efficient Learning Equilibrium}},
author = {Brafman, Ronen I. and Tennenholtz, Moshe},
booktitle = {Neural Information Processing Systems},
year = {2002},
pages = {1635-1642},
url = {https://mlanthology.org/neurips/2002/brafman2002neurips-efficient/}
}