Learning to Take Concurrent Actions
Abstract
We investigate a general semi-Markov Decision Process (SMDP) framework for modeling concurrent decision making, where agents learn optimal plans over concurrent temporally extended actions. We introduce three types of parallel termination schemes { all, any and continue { and theoretically and experimentally compare them.
Cite
Text
Rohanimanesh and Mahadevan. "Learning to Take Concurrent Actions." Neural Information Processing Systems, 2002.Markdown
[Rohanimanesh and Mahadevan. "Learning to Take Concurrent Actions." Neural Information Processing Systems, 2002.](https://mlanthology.org/neurips/2002/rohanimanesh2002neurips-learning/)BibTeX
@inproceedings{rohanimanesh2002neurips-learning,
title = {{Learning to Take Concurrent Actions}},
author = {Rohanimanesh, Khashayar and Mahadevan, Sridhar},
booktitle = {Neural Information Processing Systems},
year = {2002},
pages = {1651-1658},
url = {https://mlanthology.org/neurips/2002/rohanimanesh2002neurips-learning/}
}