Towards Exploiting Duality in Approximate Linear Programming for MDPs

Abstract

A weakness of classical methods for solving Markov decision processes is that they scale very poorly because of the flat state space, which subjects them to the curse of dimensionality. Fortunately, many MDPs are well-structured, which makes it possible to avoid enumerating the state space. To this end, factored MDP representations have been proposed (Boutilier, Dearden, & Goldszmidt 1995; Koller & Parr 1999) that model the state space as a cross product of state features, represent the transition function as a Bayesian network, and assume the rewards can be expressed as sums of compact functions of the state features. A challenge in creating algorithms for the factored representations is that well-structured problems do not always lead to compact and well-structured solutions (Koller & Parr 1999); that is, an optimal policy does not, in general, retain

Cite

Text

Dolgov and Durfee. "Towards Exploiting Duality in Approximate Linear Programming for MDPs." AAAI Conference on Artificial Intelligence, 2005. doi:10.1016/j.ajem.2017.07.057

Markdown

[Dolgov and Durfee. "Towards Exploiting Duality in Approximate Linear Programming for MDPs." AAAI Conference on Artificial Intelligence, 2005.](https://mlanthology.org/aaai/2005/dolgov2005aaai-exploiting/) doi:10.1016/j.ajem.2017.07.057

BibTeX

@inproceedings{dolgov2005aaai-exploiting,
  title     = {{Towards Exploiting Duality in Approximate Linear Programming for MDPs}},
  author    = {Dolgov, Dmitri A. and Durfee, Edmund H.},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2005},
  pages     = {1606-1607},
  doi       = {10.1016/j.ajem.2017.07.057},
  url       = {https://mlanthology.org/aaai/2005/dolgov2005aaai-exploiting/}
}