Extending Model-Based Policy Gradients for Robots in Heteroscedastic Environments
Abstract
In this paper, we consider the problem of learning robot control policies in heteroscedastic environments, whose noise properties vary throughout a robot’s state and action space. We consider reinforcement learning algorithms that evaluate policies using learned models of the environment, and we extend this class of algorithms to capture heteroscedastic effects with two enchained Gaussian processes. We explore the capabilities and limitations of this approach, and demonstrate that it reduces model bias across a variety of simulated robotic systems.
Cite
Text
Martin and Englot. "Extending Model-Based Policy Gradients for Robots in Heteroscedastic Environments." Proceedings of the 1st Annual Conference on Robot Learning, 2017.Markdown
[Martin and Englot. "Extending Model-Based Policy Gradients for Robots in Heteroscedastic Environments." Proceedings of the 1st Annual Conference on Robot Learning, 2017.](https://mlanthology.org/corl/2017/martin2017corl-extending/)BibTeX
@inproceedings{martin2017corl-extending,
title = {{Extending Model-Based Policy Gradients for Robots in Heteroscedastic Environments}},
author = {Martin, John and Englot, Brendan},
booktitle = {Proceedings of the 1st Annual Conference on Robot Learning},
year = {2017},
pages = {438-447},
volume = {78},
url = {https://mlanthology.org/corl/2017/martin2017corl-extending/}
}