A Causal World Model Underlying Next Token Prediction: Exploring GPT in a Controlled Environment

Abstract

Are generative pre-trained transformer (GPT) models, trained only to predict the next token, implicitly learning a world model from which sequences are generated one token at a time? We address this question by deriving a causal interpretation of the attention mechanism in GPT and presenting a causal world model that arises from this interpretation. Furthermore, we propose that GPT models, at inference time, can be utilized for zero-shot causal structure learning for input sequences, and introduce a corresponding confidence score. Empirical tests were conducted in controlled environments using the setups of the Othello and Chess strategy games. A GPT, pre-trained on real-world games played with the intention of winning, was tested on out-of-distribution synthetic data consisting of sequences of random legal moves. We find that the GPT model is likely to generate legal next moves for out-of-distribution sequences for which a causal structure is encoded in the attention mechanism with high confidence. In cases where it generates illegal moves, it also fails to capture a causal structure.

Cite

Text

Yehezkel Rohekar et al. "A Causal World Model Underlying Next Token Prediction: Exploring GPT in a Controlled Environment." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[Yehezkel Rohekar et al. "A Causal World Model Underlying Next Token Prediction: Exploring GPT in a Controlled Environment." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/yehezkelrohekar2025icml-causal/)

BibTeX

@inproceedings{yehezkelrohekar2025icml-causal,
  title     = {{A Causal World Model Underlying Next Token Prediction: Exploring GPT in a Controlled Environment}},
  author    = {Yehezkel Rohekar, Raanan and Gurwicz, Yaniv and Yu, Sungduk and Aflalo, Estelle and Lal, Vasudev},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {72196-72209},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/yehezkelrohekar2025icml-causal/}
}