Online Resource Allocation Using Decompositional Reinforcement Learning
Abstract
This paper considers a novel application domain for rein-forcement learning: that of “autonomic computing, ” i.e. self-managing computing systems. RL is applied to an online re-source allocation task in a distributed multi-application com-puting environment with independent time-varying load in each application. The task is to allocate servers in real time so as to maximize the sum of performance-based expected utility in each application. This task may be treated as a com-posite MDP, and to exploit the problem structure, a simple lo-calized RL approach is proposed, with better scalability than previous approaches. The RL approach is tested in a realistic prototype data center comprising real servers, real HTTP re-quests, and realistic time-varying demand. This domain poses a number of major challenges associated with live training in a real system, including: the need for rapid training, explo-ration that avoids excessive penalties, and handling complex, potentially non-Markovian system effects. The early results are encouraging: in overnight training, RL performs as well as or slightly better than heavily researched model-based ap-proaches derived from queuing theory.
Cite
Text
Tesauro. "Online Resource Allocation Using Decompositional Reinforcement Learning." AAAI Conference on Artificial Intelligence, 2005.Markdown
[Tesauro. "Online Resource Allocation Using Decompositional Reinforcement Learning." AAAI Conference on Artificial Intelligence, 2005.](https://mlanthology.org/aaai/2005/tesauro2005aaai-online/)BibTeX
@inproceedings{tesauro2005aaai-online,
title = {{Online Resource Allocation Using Decompositional Reinforcement Learning}},
author = {Tesauro, Gerald},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2005},
pages = {886-891},
url = {https://mlanthology.org/aaai/2005/tesauro2005aaai-online/}
}