A General Framework for Interacting Bayes-Optimally with Self-Interested Agents Using Arbitrary Parametric Model and Model Prior
Abstract
Recent advances in Bayesian reinforcement learning (BRL) have shown that Bayes-optimality is theoretically achievable by modeling the environment's latent dynamics using Flat-Dirichlet-Multinomial (FDM) prior. In self-interested multi-agent environments, the transition dynamics are mainly controlled by the other agent's stochastic behavior for which FDM's independence and modeling assumptions do not hold. As a result, FDM does not allow the other agent's behavior to be generalized across different states nor specified using prior domain knowledge. To overcome these practical limitations of FDM, we propose a generalization of BRL to integrate the general class of parametric models and model priors, thus allowing practitioners' domain knowledge to be exploited to produce a fine-grained and compact representation of the other agent's behavior. Empirical evaluation shows that our approach outperforms existing multi-agent reinforcement learning algorithms.
Cite
Text
Hoang and Low. "A General Framework for Interacting Bayes-Optimally with Self-Interested Agents Using Arbitrary Parametric Model and Model Prior." International Joint Conference on Artificial Intelligence, 2013.Markdown
[Hoang and Low. "A General Framework for Interacting Bayes-Optimally with Self-Interested Agents Using Arbitrary Parametric Model and Model Prior." International Joint Conference on Artificial Intelligence, 2013.](https://mlanthology.org/ijcai/2013/hoang2013ijcai-general/)BibTeX
@inproceedings{hoang2013ijcai-general,
title = {{A General Framework for Interacting Bayes-Optimally with Self-Interested Agents Using Arbitrary Parametric Model and Model Prior}},
author = {Hoang, Trong Nghia and Low, Kian Hsiang},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2013},
pages = {1394-1400},
url = {https://mlanthology.org/ijcai/2013/hoang2013ijcai-general/}
}