The Lock-in Hypothesis: Stagnation by Algorithm
Abstract
The training and deployment of large language models (LLMs) induce a feedback loop: models continually learn human beliefs from data, reinforce user beliefs with generated content, reabsorb those reinforced beliefs, and then feed them back to users. This dynamic resembles an echo chamber. We hypothesize that this feedback loop entrenches the existing values and beliefs of users, leading to a loss of diversity and potentially the *lock-in* of false beliefs. We formalize this hypothesis and test empirically with agent-based LLM simulations and real-world GPT usage data. These analyses reveal sudden but sustained drops in diversity after the release of new GPT iterations, consistent with the hypothesized human-AI feedback loop.
Cite
Text
Qiu et al. "The Lock-in Hypothesis: Stagnation by Algorithm." ICLR 2025 Workshops: Bi-Align, 2025.Markdown
[Qiu et al. "The Lock-in Hypothesis: Stagnation by Algorithm." ICLR 2025 Workshops: Bi-Align, 2025.](https://mlanthology.org/iclrw/2025/qiu2025iclrw-lockin/)BibTeX
@inproceedings{qiu2025iclrw-lockin,
title = {{The Lock-in Hypothesis: Stagnation by Algorithm}},
author = {Qiu, Tianyi and He, Zhonghao and Chugh, Tejasveer and Kleiman-Weiner, Max},
booktitle = {ICLR 2025 Workshops: Bi-Align},
year = {2025},
url = {https://mlanthology.org/iclrw/2025/qiu2025iclrw-lockin/}
}