Symbolic Dynamic Programming for Continuous State and Action MDPs
Abstract
Many real-world decision-theoretic planning problemsare naturally modeled using both continuous state andaction (CSA) spaces, yet little work has provided ex-act solutions for the case of continuous actions. Inthis work, we propose a symbolic dynamic program-ming (SDP) solution to obtain the optimal closed-formvalue function and policy for CSA-MDPs with mul-tivariate continuous state and actions, discrete noise,piecewise linear dynamics, and piecewise linear (or re-stricted piecewise quadratic) reward. Our key contribu-tion over previous SDP work is to show how the contin-uous action maximization step in the dynamic program-ming backup can be evaluated optimally and symboli-cally — a task which amounts to symbolic constrainedoptimization subject to unknown state parameters; wefurther integrate this technique to work with an efficientand compact data structure for SDP — the extendedalgebraic decision diagram (XADD). We demonstrateempirical results on a didactic nonlinear planning exam-ple and two domains from operations research to showthe first automated exact solution to these problems.
Cite
Text
Zamani et al. "Symbolic Dynamic Programming for Continuous State and Action MDPs." AAAI Conference on Artificial Intelligence, 2012. doi:10.1609/AAAI.V26I1.8372Markdown
[Zamani et al. "Symbolic Dynamic Programming for Continuous State and Action MDPs." AAAI Conference on Artificial Intelligence, 2012.](https://mlanthology.org/aaai/2012/zamani2012aaai-symbolic/) doi:10.1609/AAAI.V26I1.8372BibTeX
@inproceedings{zamani2012aaai-symbolic,
title = {{Symbolic Dynamic Programming for Continuous State and Action MDPs}},
author = {Zamani, Zahra and Sanner, Scott and Fang, Cheng},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2012},
pages = {1839-1845},
doi = {10.1609/AAAI.V26I1.8372},
url = {https://mlanthology.org/aaai/2012/zamani2012aaai-symbolic/}
}