Learning Methods to Generate Good Plans: Integrating HTN Learning and Reinforcement Learning

Abstract

We consider how to learn Hierarchical Task Networks (HTNs) for planning problems in which both the quality of solution plans generated by the HTNs and the speed at which those plans are found is important. We describe an integration of HTN Learning with Reinforcement Learning to both learn methods by analyzing semantic annotations on tasks and to produce estimates of the expected values of the learned methods by performing Monte Carlo updates. We performed an experiment in which plan quality was inversely related to plan length. In two planning domains, we evaluated the planning performance of the learned methods in comparison to two state-of-the-art satisficing classical planners, FastForward and SGPlan6, and one optimal planner, HSP*. The results demonstrate that a greedy HTN planner using the learned methods was able to generate higher quality solutions than SGPlan6 in both domains and FastForward in one. Our planner, FastForward, and SGPlan6 ran in similar time, while HSP* was exponentially slower.

Cite

Text

Hogg et al. "Learning Methods to Generate Good Plans: Integrating HTN Learning and Reinforcement Learning." AAAI Conference on Artificial Intelligence, 2010. doi:10.1609/AAAI.V24I1.7571

Markdown

[Hogg et al. "Learning Methods to Generate Good Plans: Integrating HTN Learning and Reinforcement Learning." AAAI Conference on Artificial Intelligence, 2010.](https://mlanthology.org/aaai/2010/hogg2010aaai-learning/) doi:10.1609/AAAI.V24I1.7571

BibTeX

@inproceedings{hogg2010aaai-learning,
  title     = {{Learning Methods to Generate Good Plans: Integrating HTN Learning and Reinforcement Learning}},
  author    = {Hogg, Chad and Kuter, Ugur and Muñoz-Avila, Hector},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2010},
  pages     = {1530-1535},
  doi       = {10.1609/AAAI.V24I1.7571},
  url       = {https://mlanthology.org/aaai/2010/hogg2010aaai-learning/}
}