Decision Trees That Remember: Gradient-Based Learning of Recurrent Decision Trees with Memory

Abstract

Neural architectures such as Recurrent Neural Networks (RNNs), Transformers, and State-Space Models have shown great success in handling sequential data by learning temporal dependencies. Decision Trees (DTs), on the other hand, remain a widely used class of models for structured tabular data but are typically not designed to capture sequential patterns directly. Instead, DT-based approaches for time-series data often rely on feature engineering, such as manually incorporating lag features, which can be suboptimal for capturing complex temporal dependencies. To address this limitation, we introduce ReMeDe Trees, a novel recurrent decision tree architecture that integrates an internal memory mechanism, similar to RNNs, to learn long-term dependencies in sequential data. Our model learns hard, axis-aligned decision rules for both output generation and state updates, optimizing them efficiently via gradient descent. We provide a proof-of-concept study on synthetic benchmarks to demonstrate the effectiveness of our approach.

Cite

Text

Marton et al. "Decision Trees That Remember: Gradient-Based Learning of Recurrent Decision Trees with Memory." ICLR 2025 Workshops: NFAM, 2025.

Markdown

[Marton et al. "Decision Trees That Remember: Gradient-Based Learning of Recurrent Decision Trees with Memory." ICLR 2025 Workshops: NFAM, 2025.](https://mlanthology.org/iclrw/2025/marton2025iclrw-decision/)

BibTeX

@inproceedings{marton2025iclrw-decision,
  title     = {{Decision Trees That Remember: Gradient-Based Learning of Recurrent Decision Trees with Memory}},
  author    = {Marton, Sascha and Schneider, Moritz and Brinkmann, Jannik and Lüdtke, Stefan and Bartelt, Christian and Stuckenschmidt, Heiner},
  booktitle = {ICLR 2025 Workshops: NFAM},
  year      = {2025},
  url       = {https://mlanthology.org/iclrw/2025/marton2025iclrw-decision/}
}