Generative Design of Decision Tree Policies for Reinforcement Learning

Abstract

Decision trees are an attractive choice for modeling policies in control environments due to their interpretability, conciseness, and ease of implementation. However, generating performant decision trees in this context has several challenges, including the hybrid discrete-continuous nature of the search space, the variable-length nature of the trees, the existence of parent-dependent constraints, and the high computational cost of evaluating the objective function in reinforcement learning settings. Traditional methods, such as Mixed Integer Programming or Mixed Bayesian Optimization, are unsuitable for these problems due to the variable-length constrained search space and the high number of objective function evaluations required. To address these challenges, we propose to extend approaches in the field of neural combinatorial optimization to handle the hybrid discrete-continuous optimization problem of generating decision trees. Our approach demonstrates significant improvements in performance and sample efficiency over the state-of-the-art methods for interpretable reinforcement learning with decision trees.

Cite

Text

Pettit et al. "Generative Design of Decision Tree Policies for Reinforcement Learning." ICML 2024 Workshops: SPIGM, 2024.

Markdown

[Pettit et al. "Generative Design of Decision Tree Policies for Reinforcement Learning." ICML 2024 Workshops: SPIGM, 2024.](https://mlanthology.org/icmlw/2024/pettit2024icmlw-generative/)

BibTeX

@inproceedings{pettit2024icmlw-generative,
  title     = {{Generative Design of Decision Tree Policies for Reinforcement Learning}},
  author    = {Pettit, Jacob and Lee, Chak Shing and Yang, Jiachen and Ho, Alex and Faissol, Daniel and Petersen, Brenden K. and Landajuela, Mikel},
  booktitle = {ICML 2024 Workshops: SPIGM},
  year      = {2024},
  url       = {https://mlanthology.org/icmlw/2024/pettit2024icmlw-generative/}
}