Algebraic Markov Decision Processes
Abstract
In this paper, we provide an algebraic approach to Markov Decision Processes (MDPs), which allows a unified treatment of MDPs and includes many existing models (quantitative or qualitative) as particular cases. In algebraic MDPs, rewards are expressed in a semiring structure, uncertainty is represented by a decomposable plausibility measure valued on a second semiring structure, and preferences over policies are represented by Generalized Expected Utility. We recast the problem of finding an optimal policy at a finite horizon as an algebraic path problem in a decision rule graph where arcs are valued by functions, which justifies the use of the Jacobi algorithm to solve algebraic Bellman equations. In order to show the potential of this general approach, we exhibit new variations of MDPs, admitting complete or partial preference structures, as well as probabilistic or possibilistic representation of uncertainty.
Cite
Text
Perny et al. "Algebraic Markov Decision Processes." International Joint Conference on Artificial Intelligence, 2005.Markdown
[Perny et al. "Algebraic Markov Decision Processes." International Joint Conference on Artificial Intelligence, 2005.](https://mlanthology.org/ijcai/2005/perny2005ijcai-algebraic/)BibTeX
@inproceedings{perny2005ijcai-algebraic,
title = {{Algebraic Markov Decision Processes}},
author = {Perny, Patrice and Spanjaard, Olivier and Weng, Paul},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2005},
pages = {1372-1377},
url = {https://mlanthology.org/ijcai/2005/perny2005ijcai-algebraic/}
}