Bayesian Reinforcement Learning with Behavioral Feedback
Abstract
In the standard reinforcement learning setting, the agent learns optimal policy solely from state transitions and rewards from the environment. We consider an extended setting where a trainer additionally provides feedback on the actions executed by the agent. This requires appropriately incorporating the feedback, even when the feedback is not necessarily accurate. In this paper, we present a Bayesian approach to this extended reinforcement learning setting. Specifically, we extend Kalman Temporal Difference learning to compute the posterior distribution over Q-values given the state transitions and rewards from the environment as well as the feedback from the trainer. Through experiments on standard reinforcement learning tasks, we show that learning performance can be significantly improved even with inaccurate feedback. PDF
Cite
Text
Hong et al. "Bayesian Reinforcement Learning with Behavioral Feedback." International Joint Conference on Artificial Intelligence, 2016.Markdown
[Hong et al. "Bayesian Reinforcement Learning with Behavioral Feedback." International Joint Conference on Artificial Intelligence, 2016.](https://mlanthology.org/ijcai/2016/hong2016ijcai-bayesian/)BibTeX
@inproceedings{hong2016ijcai-bayesian,
title = {{Bayesian Reinforcement Learning with Behavioral Feedback}},
author = {Hong, Teakgyu and Lee, Jongmin and Kim, Kee-Eung and Ortega, Pedro A. and Lee, Daniel D.},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2016},
pages = {1571-1577},
url = {https://mlanthology.org/ijcai/2016/hong2016ijcai-bayesian/}
}