Approximating Optimal Policies for Agents with Limited Execution Resources

Abstract

An agent with limited consumable execution re-sources needs policies that attempt to achieve good performance while respecting these limitations. Otherwise, an agent (such as a plane) might fail catastrophically (crash) when it runs out of re-sources (fuel) at the wrong time (in midair). We present a new approach to constructing policies for agents with limited execution resources that builds on principles of real-time AI, as well as research in constrained Markov decision processes. Specif-ically, we formulate, solve, and analyze the pol-icy optimization problem where constraints are im-posed on the probability of exceeding the resource limits. We describe and empirically evaluate our solution technique to show that it is computation-ally reasonable, and that it generates policies that sacrifice some potential reward in order to make the kinds of precise guarantees about the probability of resource overutilization that are crucial for mission-critical applications. 1

Cite

Text

Dolgov and Durfee. "Approximating Optimal Policies for Agents with Limited Execution Resources." International Joint Conference on Artificial Intelligence, 2003.

Markdown

[Dolgov and Durfee. "Approximating Optimal Policies for Agents with Limited Execution Resources." International Joint Conference on Artificial Intelligence, 2003.](https://mlanthology.org/ijcai/2003/dolgov2003ijcai-approximating/)

BibTeX

@inproceedings{dolgov2003ijcai-approximating,
  title     = {{Approximating Optimal Policies for Agents with Limited Execution Resources}},
  author    = {Dolgov, Dmitri A. and Durfee, Edmund H.},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2003},
  pages     = {1107-1112},
  url       = {https://mlanthology.org/ijcai/2003/dolgov2003ijcai-approximating/}
}