Intention Progression with Temporally Extended Goals

Abstract

This paper introduces a novel method for estimating the self-interest level of Markov social dilemmas. We extend the concept of self-interest level from normal-form games to Markov games, providing a quantitative measure of the minimum reward exchange required to align individual and collective interests. We demonstrate our method on three environments from the Melting Pot suite, representing either common-pool resources or public goods. Our results illustrate how reward exchange can enable agents to transition from selfish to collective equilibria in a Markov social dilemma. This work contributes to multi-agent reinforcement learning by providing a practical tool for analysing complex, multistep social dilemmas. Our findings offer insights into how reward structures can promote or hinder cooperation, with potential applications in areas such as mechanism design.

Cite

Text

Yao et al. "Intention Progression with Temporally Extended Goals." International Joint Conference on Artificial Intelligence, 2024. doi:10.24963/ijcai.2024/33

Markdown

[Yao et al. "Intention Progression with Temporally Extended Goals." International Joint Conference on Artificial Intelligence, 2024.](https://mlanthology.org/ijcai/2024/yao2024ijcai-intention/) doi:10.24963/ijcai.2024/33

BibTeX

@inproceedings{yao2024ijcai-intention,
  title     = {{Intention Progression with Temporally Extended Goals}},
  author    = {Yao, Yuan and Alechina, Natasha and Logan, Brian},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2024},
  pages     = {292-301},
  doi       = {10.24963/ijcai.2024/33},
  url       = {https://mlanthology.org/ijcai/2024/yao2024ijcai-intention/}
}