Point-Based Value Iteration for Constrained POMDPs
Abstract
Constrained partially observable Markov decision processes (CPOMDPs) extend the standard POMDPs by allowing the specification of constraints on some aspects of the policy in addition to the optimality objective for the value function. CPOMDPs have many practical advantages over standard POMDPs since they naturally model problems involving limited resource or multiple objectives. In this paper, we show that the optimal policies in CPOMDPs can be randomized, and present exact and approximate dynamic programming methods for computing randomized optimal policies. While the exact method requires solving a minimax quadratically constrained program (QCP) in each dynamic programming update, the approximate method utilizes the point-based value update with a linear program (LP). We show that the randomized policies are significantly better than the deterministic ones. We also demonstrate that the approximate point-based method is scalable to solve large problems.
Cite
Text
Kim et al. "Point-Based Value Iteration for Constrained POMDPs." International Joint Conference on Artificial Intelligence, 2011. doi:10.5591/978-1-57735-516-8/IJCAI11-329Markdown
[Kim et al. "Point-Based Value Iteration for Constrained POMDPs." International Joint Conference on Artificial Intelligence, 2011.](https://mlanthology.org/ijcai/2011/kim2011ijcai-point/) doi:10.5591/978-1-57735-516-8/IJCAI11-329BibTeX
@inproceedings{kim2011ijcai-point,
title = {{Point-Based Value Iteration for Constrained POMDPs}},
author = {Kim, Dongho and Lee, Jaesong and Kim, Kee-Eung and Poupart, Pascal},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2011},
pages = {1968-1974},
doi = {10.5591/978-1-57735-516-8/IJCAI11-329},
url = {https://mlanthology.org/ijcai/2011/kim2011ijcai-point/}
}