On Stochastic Optimal Control and Reinforcement Learning by Approximate Inference (Extended Abstract)

Abstract

We present a reformulation of the stochastic optimal control problem in terms of KL divergence minimisation, not only providing a unifying perspective of previous approaches in this area, but also demonstrating that the formalism leads to novel practical approaches to the control problem. Specifically, a natural relaxation of the dual formulation gives rise to exact iterative solutions to the finite and infinite horizon stochastic optimal control problem, while direct application of Bayesian inference methods yields instances of risk sensitive control.

Cite

Text

Rawlik et al. "On Stochastic Optimal Control and Reinforcement Learning by Approximate Inference (Extended Abstract)." International Joint Conference on Artificial Intelligence, 2013.

Markdown

[Rawlik et al. "On Stochastic Optimal Control and Reinforcement Learning by Approximate Inference (Extended Abstract)." International Joint Conference on Artificial Intelligence, 2013.](https://mlanthology.org/ijcai/2013/rawlik2013ijcai-stochastic/)

BibTeX

@inproceedings{rawlik2013ijcai-stochastic,
  title     = {{On Stochastic Optimal Control and Reinforcement Learning by Approximate Inference (Extended Abstract)}},
  author    = {Rawlik, Konrad and Toussaint, Marc and Vijayakumar, Sethu},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2013},
  pages     = {3052-3056},
  url       = {https://mlanthology.org/ijcai/2013/rawlik2013ijcai-stochastic/}
}