Ants and Reinforcement Learning: A Case Study in Routing in Dynamic Networks
Abstract
We investigate two new distributed routing algorithms for data networks based on simple biological "ants" that explore the network and rapidly learn good routes, using a novel variation of reinforcement learning. These two algorithms are fully adaptive to topology changes and changes in link costs in the network, and have space and computational overheads that are competitive with traditional packet routing algorithms: although they can generate more routing traffic when the rate of failures in a network is low, they perform much better under higher failure rates. Both algorithms are more resilient than traditional algorithms, in the sense that random corruption of routing state has limited impact on the computation of paths. We present convergence theorems for both of our algorithms drawing on the theory of non-stationary and stationary discrete-time Markov chains over the reals. We present an extensive empirical evaluation of our algorithms on a simulator that is widely used in the c...
Cite
Text
Subramanian et al. "Ants and Reinforcement Learning: A Case Study in Routing in Dynamic Networks." International Joint Conference on Artificial Intelligence, 1997.Markdown
[Subramanian et al. "Ants and Reinforcement Learning: A Case Study in Routing in Dynamic Networks." International Joint Conference on Artificial Intelligence, 1997.](https://mlanthology.org/ijcai/1997/subramanian1997ijcai-ants/)BibTeX
@inproceedings{subramanian1997ijcai-ants,
title = {{Ants and Reinforcement Learning: A Case Study in Routing in Dynamic Networks}},
author = {Subramanian, Devika and Druschel, Peter and Chen, Johnny},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {1997},
pages = {832-839},
url = {https://mlanthology.org/ijcai/1997/subramanian1997ijcai-ants/}
}