Sparse Reinforcement Learning via Convex Optimization

Abstract

We propose two new algorithms for the sparse reinforcement learning problem based on different formulations. The first algorithm is an off-line method based on the alternating direction method of multipliers for solving a constrained formulation that explicitly controls the projected Bellman residual. The second algorithm is an online stochastic approximation algorithm that employs the regularized dual averaging technique, using the Lagrangian formulation. The convergence of both algorithms are established. We demonstrate the performance of these algorithms through two classical examples.

Cite

Text

Qin et al. "Sparse Reinforcement Learning via Convex Optimization." International Conference on Machine Learning, 2014.

Markdown

[Qin et al. "Sparse Reinforcement Learning via Convex Optimization." International Conference on Machine Learning, 2014.](https://mlanthology.org/icml/2014/qin2014icml-sparse/)

BibTeX

@inproceedings{qin2014icml-sparse,
  title     = {{Sparse Reinforcement Learning via Convex Optimization}},
  author    = {Qin, Zhiwei and Li, Weichang and Janoos, Firdaus},
  booktitle = {International Conference on Machine Learning},
  year      = {2014},
  pages     = {424-432},
  volume    = {32},
  url       = {https://mlanthology.org/icml/2014/qin2014icml-sparse/}
}