Efficient Duple Perturbation Robustness in Low-Rank MDPs

Abstract

The pursuit of robustness has recently been a popular topic in reinforcement learning (RL) research, yet the existing methods generally suffer from computation issues that obstruct their real-world implementation. In this paper, we consider MDPs with low-rank structures, where the transition kernel can be written as a linear product of feature map and factors. We introduce *duple perturbation* robustness, i.e. perturbation on both the feature map and the factors, via a novel characterization of $(\xi,\eta)$-ambiguity sets featuring computational efficiency. Our novel low-rank robust MDP formulation is compatible with the low-rank function representation view, and therefore, is naturally applicable to practical RL problems with large or even continuous state-action spaces. Meanwhile, it also gives rise to a provably efficient and practical algorithm with theoretical convergence rate guarantee. Lastly, the robustness of our proposed approach is justified by numerical experiments, including classical control tasks with continuous state-action spaces.

Cite

Text

Hu et al. "Efficient Duple Perturbation Robustness in Low-Rank MDPs." Proceedings of the 7th Annual Learning for Dynamics \& Control Conference, 2025.

Markdown

[Hu et al. "Efficient Duple Perturbation Robustness in Low-Rank MDPs." Proceedings of the 7th Annual Learning for Dynamics \& Control Conference, 2025.](https://mlanthology.org/l4dc/2025/hu2025l4dc-efficient/)

BibTeX

@inproceedings{hu2025l4dc-efficient,
  title     = {{Efficient Duple Perturbation Robustness in Low-Rank MDPs}},
  author    = {Hu, Yang and Ma, Haitong and Li, Na and Dai, Bo},
  booktitle = {Proceedings of the 7th Annual Learning for Dynamics \& Control Conference},
  year      = {2025},
  pages     = {723-737},
  volume    = {283},
  url       = {https://mlanthology.org/l4dc/2025/hu2025l4dc-efficient/}
}