Density Constrained Reinforcement Learning

Abstract

We study constrained reinforcement learning (CRL) from a novel perspective by setting constraints directly on state density functions, rather than the value functions considered by previous works. State density has a clear physical and mathematical interpretation, and is able to express a wide variety of constraints such as resource limits and safety requirements. Density constraints can also avoid the time-consuming process of designing and tuning cost functions required by value function-based constraints to encode system specifications. We leverage the duality between density functions and Q functions to develop an effective algorithm to solve the density constrained RL problem optimally and the constrains are guaranteed to be satisfied. We prove that the proposed algorithm converges to a near-optimal solution with a bounded error even when the policy update is imperfect. We use a set of comprehensive experiments to demonstrate the advantages of our approach over state-of-the-art CRL methods, with a wide range of density constrained tasks as well as standard CRL benchmarks such as Safety-Gym.

Cite

Text

Qin et al. "Density Constrained Reinforcement Learning." International Conference on Machine Learning, 2021.

Markdown

[Qin et al. "Density Constrained Reinforcement Learning." International Conference on Machine Learning, 2021.](https://mlanthology.org/icml/2021/qin2021icml-density/)

BibTeX

@inproceedings{qin2021icml-density,
  title     = {{Density Constrained Reinforcement Learning}},
  author    = {Qin, Zengyi and Chen, Yuxiao and Fan, Chuchu},
  booktitle = {International Conference on Machine Learning},
  year      = {2021},
  pages     = {8682-8692},
  volume    = {139},
  url       = {https://mlanthology.org/icml/2021/qin2021icml-density/}
}