Safeguarded Progress in Reinforcement Learning: Safe Bayesian Exploration for Control Policy Synthesis

Abstract

This paper addresses the problem of maintaining safety during training in Reinforcement Learning (RL), such that the safety constraint violations are bounded at any point during learning. As enforcing safety during training might severely limit the agent’s exploration, we propose here a new architecture that handles the trade-off between efficient progress and safety during exploration. As the exploration progresses, we update via Bayesian inference Dirichlet-Categorical models of the transition probabilities of the Markov decision process that describes the environment dynamics. We then propose a way to approximate moments of belief about the risk associated to the action selection policy. We demonstrate that this approach can be easily interleaved with RL and we present experimental results to showcase the performance of the overall architecture.

Cite

Text

Mitta et al. "Safeguarded Progress in Reinforcement Learning: Safe Bayesian Exploration for Control Policy Synthesis." AAAI Conference on Artificial Intelligence, 2024. doi:10.1609/AAAI.V38I19.30137

Markdown

[Mitta et al. "Safeguarded Progress in Reinforcement Learning: Safe Bayesian Exploration for Control Policy Synthesis." AAAI Conference on Artificial Intelligence, 2024.](https://mlanthology.org/aaai/2024/mitta2024aaai-safeguarded/) doi:10.1609/AAAI.V38I19.30137

BibTeX

@inproceedings{mitta2024aaai-safeguarded,
  title     = {{Safeguarded Progress in Reinforcement Learning: Safe Bayesian Exploration for Control Policy Synthesis}},
  author    = {Mitta, Rohan and Hasanbeig, Hosein and Wang, Jun and Kroening, Daniel and Kantaros, Yiannis and Abate, Alessandro},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2024},
  pages     = {21412-21419},
  doi       = {10.1609/AAAI.V38I19.30137},
  url       = {https://mlanthology.org/aaai/2024/mitta2024aaai-safeguarded/}
}