Value of Information and Reward Specification in Active Inference and POMDPs

Abstract

Active inference is an agent modeling framework with roots in Bayesian predictive coding and the free energy principle. In recent years, active inference has gained popularity in modeling sequential decision making, a problem set traditionally populated by reinforcement learning (RL). Instead of optimizing expected reward as in RL, active inference agents optimize expected free energy (EFE), which has an intuitive decomposition into a pragmatic and an epistemic component. This makes us wonder: what's the EFE-optimizing agent's optimality gap compared with a reward-driven RL agent, which is well understood? By casting EFE under a particular class of belief MDP and using analysis tools from RL theory, we show that EFE approximates the Bayes optimal RL policy via information value. We discuss the implications for objective specification of active inference agents.

Cite

Text

Wei. "Value of Information and Reward Specification in Active Inference and POMDPs." NeurIPS 2024 Workshops: NeuroAI, 2024.

Markdown

[Wei. "Value of Information and Reward Specification in Active Inference and POMDPs." NeurIPS 2024 Workshops: NeuroAI, 2024.](https://mlanthology.org/neuripsw/2024/wei2024neuripsw-value/)

BibTeX

@inproceedings{wei2024neuripsw-value,
  title     = {{Value of Information and Reward Specification in Active Inference and POMDPs}},
  author    = {Wei, Ran},
  booktitle = {NeurIPS 2024 Workshops: NeuroAI},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/wei2024neuripsw-value/}
}