Advice Generation from Observed Execution: Abstract Markov Decision Process Learning

Abstract

An advising agent, a coach, provides advice to other agents about how to act. In this paper we contribute an advice generation method using observations of agents acting in an environment. Given an abstract state definition and partially specified abstract actions, the algorithm extracts a Markov Chain, infers a Markov Decision Process, and then solves the MDP (given an arbitrary reward signal) to generate advice. We evaluate our work in a simulated robot soccer environment and experimental results show improved agent performance when using the advice generated from the MDP for both a sub-task and the full soccer game.

Cite

Text

Riley and Veloso. "Advice Generation from Observed Execution: Abstract Markov Decision Process Learning." AAAI Conference on Artificial Intelligence, 2004.

Markdown

[Riley and Veloso. "Advice Generation from Observed Execution: Abstract Markov Decision Process Learning." AAAI Conference on Artificial Intelligence, 2004.](https://mlanthology.org/aaai/2004/riley2004aaai-advice/)

BibTeX

@inproceedings{riley2004aaai-advice,
  title     = {{Advice Generation from Observed Execution: Abstract Markov Decision Process Learning}},
  author    = {Riley, Patrick and Veloso, Manuela M.},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2004},
  pages     = {631-637},
  url       = {https://mlanthology.org/aaai/2004/riley2004aaai-advice/}
}