Multi-Agent Reinforcement Learning: Independent Versus Cooperative Agents
Abstract
Intelligent human agents exist in a cooperative social environment that facilitates learning. They learn not only by trial-and-error, but also through cooperation by sharing instantaneous information, episodic experience, and learned knowledge. The key investigations of this paper are, “Given the same number of reinforcement learning agents, will cooperative agents outperform independent agents who do not communicate during learning?” and “What is the price for such cooperation?” Using independent agents as a benchmark, cooperative agents are studied in following ways: (1) sharing sensation, (2) sharing episodes, and (3) sharing learned policies. This paper shows that (a) additional sensation from another agent is beneficial if it can be used efficiently, (b) sharing learned policies or episodes among agents speeds up learning at the cost of communication, and (c) for joint tasks, agents engaging in partnership can significantly outperform independent agents although they may learn slowly in the beginning. These tradeoff's are not just limited to multi-agent reinforcement learning.
Cite
Text
Tan. "Multi-Agent Reinforcement Learning: Independent Versus Cooperative Agents." International Conference on Machine Learning, 1993. doi:10.1016/B978-1-55860-307-3.50049-6Markdown
[Tan. "Multi-Agent Reinforcement Learning: Independent Versus Cooperative Agents." International Conference on Machine Learning, 1993.](https://mlanthology.org/icml/1993/tan1993icml-multi/) doi:10.1016/B978-1-55860-307-3.50049-6BibTeX
@inproceedings{tan1993icml-multi,
title = {{Multi-Agent Reinforcement Learning: Independent Versus Cooperative Agents}},
author = {Tan, Ming},
booktitle = {International Conference on Machine Learning},
year = {1993},
pages = {330-337},
doi = {10.1016/B978-1-55860-307-3.50049-6},
url = {https://mlanthology.org/icml/1993/tan1993icml-multi/}
}