Robust Algorithmic Collusion
Abstract
This paper develops an approach to assess reinforcement learners with collusive pricing policies in a testing environment. We find that algorithms are unable to extrapolate collusive policies from their training environment to testing environments. Collusion consistently breaks down, and algorithms instead tend to converge to Nash prices. Policy updating with or without exploration re-establishes collusion, but only in the current environment. This is robust to repeated learning across environments. Our results indicate that frequent market interaction, coordination of algorithm design, and stable environments are essential for algorithmic collusion.
Cite
Text
Eschenbaum and Zahn. "Robust Algorithmic Collusion." NeurIPS 2021 Workshops: PERLS, 2021.Markdown
[Eschenbaum and Zahn. "Robust Algorithmic Collusion." NeurIPS 2021 Workshops: PERLS, 2021.](https://mlanthology.org/neuripsw/2021/eschenbaum2021neuripsw-robust/)BibTeX
@inproceedings{eschenbaum2021neuripsw-robust,
title = {{Robust Algorithmic Collusion}},
author = {Eschenbaum, Nicolas and Zahn, Philipp},
booktitle = {NeurIPS 2021 Workshops: PERLS},
year = {2021},
url = {https://mlanthology.org/neuripsw/2021/eschenbaum2021neuripsw-robust/}
}