Learning Optimal Decision Trees Under Memory Constraints

Abstract

Existing algorithms for learning optimal decision trees can be put into two categories: algorithms based on the use of Mixed Integer Programming (MIP) solvers and algorithms based on dynamic programming (DP) on itemsets. While the algorithms based on DP are the fastest, their main disadvantage compared to MIP-based approaches is that the amount of memory these algorithms may require to find an optimal solution is not bounded. Consequently, for some datasets these algorithms can only be executed on machines with large amounts of memory. In this paper, we propose the first DP-based algorithm for learning optimal decision trees that operates under memory constraints. Core contributions of this work include: (1) strategies for freeing memory when too much memory is used by the algorithm; (2) an effective approach for recovering the optimal decision tree when parts of the memory are freed. Our experiments demonstrate a favorable trade-off between memory constraints and the run times of our algorithm.

Cite

Text

Aglin et al. "Learning Optimal Decision Trees Under Memory Constraints." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2022. doi:10.1007/978-3-031-26419-1_24

Markdown

[Aglin et al. "Learning Optimal Decision Trees Under Memory Constraints." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2022.](https://mlanthology.org/ecmlpkdd/2022/aglin2022ecmlpkdd-learning/) doi:10.1007/978-3-031-26419-1_24

BibTeX

@inproceedings{aglin2022ecmlpkdd-learning,
  title     = {{Learning Optimal Decision Trees Under Memory Constraints}},
  author    = {Aglin, Gaël and Nijssen, Siegfried and Schaus, Pierre},
  booktitle = {European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases},
  year      = {2022},
  pages     = {393-409},
  doi       = {10.1007/978-3-031-26419-1_24},
  url       = {https://mlanthology.org/ecmlpkdd/2022/aglin2022ecmlpkdd-learning/}
}