Regularized Feature Selection in Reinforcement Learning

Abstract

We introduce feature regularization during feature selection for value function approximation. Feature regularization introduces a prior into the selection process, improving function approximation accuracy and reducing overfitting. We show that the smoothness prior is effective in the incremental feature selection setting and present closed-form smoothness regularizers for the Fourier and RBF bases. We present two methods for feature regularization which extend the temporal difference orthogonal matching pursuit (OMP-TD) algorithm and demonstrate the effectiveness of the smoothness prior; smooth Tikhonov OMP-TD and smoothness scaled OMP-TD. We compare these methods against OMP-TD, regularized OMP-TD and least squares TD with random projections, across six benchmark domains using two different types of basis functions.

Cite

Text

Wookey and Konidaris. "Regularized Feature Selection in Reinforcement Learning." Machine Learning, 2015. doi:10.1007/S10994-015-5518-8

Markdown

[Wookey and Konidaris. "Regularized Feature Selection in Reinforcement Learning." Machine Learning, 2015.](https://mlanthology.org/mlj/2015/wookey2015mlj-regularized/) doi:10.1007/S10994-015-5518-8

BibTeX

@article{wookey2015mlj-regularized,
  title     = {{Regularized Feature Selection in Reinforcement Learning}},
  author    = {Wookey, Dean S. and Konidaris, George Dimitri},
  journal   = {Machine Learning},
  year      = {2015},
  pages     = {655-676},
  doi       = {10.1007/S10994-015-5518-8},
  volume    = {100},
  url       = {https://mlanthology.org/mlj/2015/wookey2015mlj-regularized/}
}