Learning Small Decision Trees with Few Outliers: A Parameterized Perspective
Abstract
Decision trees is a fundamental tool in machine learning for representing, classifying, and generalizing data. It is desirable to construct ``small'' decision trees, by minimizing either the size (s) or the depth (d) of the decision tree (DT). Recently, the parameterized complexity of Decision Tree Learning has attracted a lot of attention. We consider a generalization of Decision Tree Learning where given a classification instance E and an integer t, the task is to find a ``small'' DT that disagrees with E in at most t examples. We consider two problems: DTSO and DTDO, where the goal is to construct a DT minimizing s and d, respectively. We first establish that both DTSO and DTDO are W[1]-hard when parameterized by s+y and d+y, respectively, where y is the maximum number of features in which two differently labeled examples can differ. We complement this result by showing that these problems become FPT if we include the parameter t. We also consider the kernelization complexity of these problems and establish several positive and negative results for both DTSO and DTDO.
Cite
Text
Gahlawat and Zehavi. "Learning Small Decision Trees with Few Outliers: A Parameterized Perspective." AAAI Conference on Artificial Intelligence, 2024. doi:10.1609/AAAI.V38I11.29098Markdown
[Gahlawat and Zehavi. "Learning Small Decision Trees with Few Outliers: A Parameterized Perspective." AAAI Conference on Artificial Intelligence, 2024.](https://mlanthology.org/aaai/2024/gahlawat2024aaai-learning/) doi:10.1609/AAAI.V38I11.29098BibTeX
@inproceedings{gahlawat2024aaai-learning,
title = {{Learning Small Decision Trees with Few Outliers: A Parameterized Perspective}},
author = {Gahlawat, Harmender and Zehavi, Meirav},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2024},
pages = {12100-12108},
doi = {10.1609/AAAI.V38I11.29098},
url = {https://mlanthology.org/aaai/2024/gahlawat2024aaai-learning/}
}