Instance-Based Online Learning of Deterministic Relational Action Models

Abstract

We present an instance-based, online method for learning action models in unanticipated, relational domains. Our algorithm memorizes pre- and post-states of transitions an agent encounters while experiencing the environment, and makes predictions by using analogy to map the recorded transitions to novel situations. Our algorithm is implemented in the Soar cognitive architecture, integrating its task-independent episodic memory module and analogical reasoning implemented in procedural memory. We evaluate this algorithm’s prediction performance in a modified version of the blocks world domain and the taxi domain. We also present a reinforcement learning agent that uses our model learning algorithm to significantly speed up its convergence to an optimal policy in the modified blocks world domain.

Cite

Text

Xu and Laird. "Instance-Based Online Learning of Deterministic Relational Action Models." AAAI Conference on Artificial Intelligence, 2010. doi:10.1609/AAAI.V24I1.7569

Markdown

[Xu and Laird. "Instance-Based Online Learning of Deterministic Relational Action Models." AAAI Conference on Artificial Intelligence, 2010.](https://mlanthology.org/aaai/2010/xu2010aaai-instance/) doi:10.1609/AAAI.V24I1.7569

BibTeX

@inproceedings{xu2010aaai-instance,
  title     = {{Instance-Based Online Learning of Deterministic Relational Action Models}},
  author    = {Xu, Joseph Z. and Laird, John E.},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2010},
  pages     = {1574-1579},
  doi       = {10.1609/AAAI.V24I1.7569},
  url       = {https://mlanthology.org/aaai/2010/xu2010aaai-instance/}
}