Alternating Optimisation and Quadrature for Robust Control

Abstract

Bayesian optimisation has been successfully applied to a variety of reinforcement learning problems. However, the traditional approach for learning optimal policies in simulators does not utilise the opportunity to improve learning by adjusting certain environment variables: state features that are unobservable and randomly determined by the environment in a physical setting but are controllable in a simulator. This paper considers the problem of finding a robust policy while taking into account the impact of environment variables. We present Alternating Optimisation and Quadrature (ALOQ), which uses Bayesian optimisation and Bayesian quadrature to address such settings. ALOQ is robust to the presence of significant rare events, which may not be observable under random sampling, but play a substantial role in determining the optimal policy. Experimental results across different domains show that ALOQ can learn more efficiently and robustly than existing methods.

Cite

Text

Paul et al. "Alternating Optimisation and Quadrature for Robust Control." AAAI Conference on Artificial Intelligence, 2018. doi:10.1609/AAAI.V32I1.11687

Markdown

[Paul et al. "Alternating Optimisation and Quadrature for Robust Control." AAAI Conference on Artificial Intelligence, 2018.](https://mlanthology.org/aaai/2018/paul2018aaai-alternating/) doi:10.1609/AAAI.V32I1.11687

BibTeX

@inproceedings{paul2018aaai-alternating,
  title     = {{Alternating Optimisation and Quadrature for Robust Control}},
  author    = {Paul, Supratik and Chatzilygeroudis, Konstantinos I. and Ciosek, Kamil and Mouret, Jean-Baptiste and Osborne, Michael A. and Whiteson, Shimon},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2018},
  pages     = {3925-3933},
  doi       = {10.1609/AAAI.V32I1.11687},
  url       = {https://mlanthology.org/aaai/2018/paul2018aaai-alternating/}
}