Hard Constrained Semi-Markov Decision Processes
Abstract
In multiple criteria Markov Decision Processes (MDP) where multiple costs are incurred at every decision point, current methods solve them by minimising the expected primary cost criterion while constraining the expectations of other cost criteria to some critical values. However, systems are of-ten faced with hard constraints where the cost criteria should never exceed some critical values at any time, rather than con-straints based on the expected cost criteria. For example, a resource-limited sensor network no longer functions once its energy is depleted. Based on the semi-MDP (sMDP) model, we study the hard constrained (HC) problem in continuous time, state and action spaces with respect to both finite and infinite horizons, and various cost criteria. We show that the HCsMDP problem is NP-hard and that there exists an equiv-alent discrete-time MDP to every HCsMDP. Hence, classical methods such as reinforcement learning can solve HCsMDPs.
Cite
Text
Yeow et al. "Hard Constrained Semi-Markov Decision Processes." AAAI Conference on Artificial Intelligence, 2006.Markdown
[Yeow et al. "Hard Constrained Semi-Markov Decision Processes." AAAI Conference on Artificial Intelligence, 2006.](https://mlanthology.org/aaai/2006/yeow2006aaai-hard/)BibTeX
@inproceedings{yeow2006aaai-hard,
title = {{Hard Constrained Semi-Markov Decision Processes}},
author = {Yeow, Wai-Leong and Tham, Chen-Khong and Wong, Wai-Choong},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2006},
pages = {549-554},
url = {https://mlanthology.org/aaai/2006/yeow2006aaai-hard/}
}