Explainable Agency in Reinforcement Learning Agents
Abstract
This thesis explores how reinforcement learning (RL) agents can provide explanations for their actions and behaviours. As humans, we build causal models to encode cause-effect relations of events and use these to explain why events happen. Taking inspiration from cognitive psychology and social science literature, I build causal explanation models and explanation dialogue models for RL agents. By mimicking human-like explanation models, these agents can provide explanations that are natural and intuitive to humans.
Cite
Text
Madumal. "Explainable Agency in Reinforcement Learning Agents." AAAI Conference on Artificial Intelligence, 2020. doi:10.1609/AAAI.V34I10.7134Markdown
[Madumal. "Explainable Agency in Reinforcement Learning Agents." AAAI Conference on Artificial Intelligence, 2020.](https://mlanthology.org/aaai/2020/madumal2020aaai-explainable-a/) doi:10.1609/AAAI.V34I10.7134BibTeX
@inproceedings{madumal2020aaai-explainable-a,
title = {{Explainable Agency in Reinforcement Learning Agents}},
author = {Madumal, Prashan},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2020},
pages = {13724-13725},
doi = {10.1609/AAAI.V34I10.7134},
url = {https://mlanthology.org/aaai/2020/madumal2020aaai-explainable-a/}
}