Instance-Based Online Learning of Deterministic Relational Action Models
Abstract
We present an instance-based, online method for learning action models in unanticipated, relational domains. Our algorithm memorizes pre- and post-states of transitions an agent encounters while experiencing the environment, and makes predictions by using analogy to map the recorded transitions to novel situations. Our algorithm is implemented in the Soar cognitive architecture, integrating its task-independent episodic memory module and analogical reasoning implemented in procedural memory. We evaluate this algorithm’s prediction performance in a modified version of the blocks world domain and the taxi domain. We also present a reinforcement learning agent that uses our model learning algorithm to significantly speed up its convergence to an optimal policy in the modified blocks world domain.
Cite
Text
Xu and Laird. "Instance-Based Online Learning of Deterministic Relational Action Models." AAAI Conference on Artificial Intelligence, 2010. doi:10.1609/AAAI.V24I1.7569Markdown
[Xu and Laird. "Instance-Based Online Learning of Deterministic Relational Action Models." AAAI Conference on Artificial Intelligence, 2010.](https://mlanthology.org/aaai/2010/xu2010aaai-instance/) doi:10.1609/AAAI.V24I1.7569BibTeX
@inproceedings{xu2010aaai-instance,
title = {{Instance-Based Online Learning of Deterministic Relational Action Models}},
author = {Xu, Joseph Z. and Laird, John E.},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2010},
pages = {1574-1579},
doi = {10.1609/AAAI.V24I1.7569},
url = {https://mlanthology.org/aaai/2010/xu2010aaai-instance/}
}