Efficient Duple Perturbation Robustness in Low-Rank MDPs
Abstract
The pursuit of robustness has recently been a popular topic in reinforcement learning (RL) research, yet the existing methods generally suffer from computation issues that obstruct their real-world implementation. In this paper, we consider MDPs with low-rank structures, where the transition kernel can be written as a linear product of feature map and factors. We introduce *duple perturbation* robustness, i.e. perturbation on both the feature map and the factors, via a novel characterization of $(\xi,\eta)$-ambiguity sets featuring computational efficiency. Our novel low-rank robust MDP formulation is compatible with the low-rank function representation view, and therefore, is naturally applicable to practical RL problems with large or even continuous state-action spaces. Meanwhile, it also gives rise to a provably efficient and practical algorithm with theoretical convergence rate guarantee. Lastly, the robustness of our proposed approach is justified by numerical experiments, including classical control tasks with continuous state-action spaces.
Cite
Text
Hu et al. "Efficient Duple Perturbation Robustness in Low-Rank MDPs." Proceedings of the 7th Annual Learning for Dynamics \& Control Conference, 2025.Markdown
[Hu et al. "Efficient Duple Perturbation Robustness in Low-Rank MDPs." Proceedings of the 7th Annual Learning for Dynamics \& Control Conference, 2025.](https://mlanthology.org/l4dc/2025/hu2025l4dc-efficient/)BibTeX
@inproceedings{hu2025l4dc-efficient,
title = {{Efficient Duple Perturbation Robustness in Low-Rank MDPs}},
author = {Hu, Yang and Ma, Haitong and Li, Na and Dai, Bo},
booktitle = {Proceedings of the 7th Annual Learning for Dynamics \& Control Conference},
year = {2025},
pages = {723-737},
volume = {283},
url = {https://mlanthology.org/l4dc/2025/hu2025l4dc-efficient/}
}