Risk-Averse Constrained Reinforcement Learning with Optimized Certainty Equivalents

Abstract

Constrained optimization provides a common framework for dealing with conflicting objectives in reinforcement learning (RL). In most of these settings, the objectives (and constraints) are expressed though the expected accumulated reward. However, this formulation neglects risky or even possibly catastrophic events at the tails of the reward distribution, and is often insufficient for high-stakes applications in which the risk involved in outliers is critical. In this work, we propose a framework for risk-aware constrained RL, which exhibits per-stage robustness properties jointly in reward values and time using optimized certainty equivalents (OCEs). Our framework ensures an exact equivalent to the original constrained problem within a parameterized strong Lagrangian duality framework under appropriate constraint qualifications, and yields a simple algorithmic recipe which can be wrapped around standard RL solvers, such as PPO. Lastly, we establish the convergence of the proposed algorithm and verify the risk-aware properties of our approach through several numerical experiments.

Cite

Text

Lee et al. "Risk-Averse Constrained Reinforcement Learning with Optimized Certainty Equivalents." Advances in Neural Information Processing Systems, 2025.

Markdown

[Lee et al. "Risk-Averse Constrained Reinforcement Learning with Optimized Certainty Equivalents." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/lee2025neurips-riskaverse/)

BibTeX

@inproceedings{lee2025neurips-riskaverse,
  title     = {{Risk-Averse Constrained Reinforcement Learning with Optimized Certainty Equivalents}},
  author    = {Lee, Jane H. and Saglam, Baturay and Pougkakiotis, Spyridon and Karbasi, Amin and Kalogerias, Dionysis},
  booktitle = {Advances in Neural Information Processing Systems},
  year      = {2025},
  url       = {https://mlanthology.org/neurips/2025/lee2025neurips-riskaverse/}
}