TY - GEN
T1 - Learning Small Decision Trees With Few Outliers
T2 - 38th AAAI Conference on Artificial Intelligence, AAAI 2024
AU - Gahlawat, Harmender
AU - Zehavi, Meirav
N1 - Publisher Copyright:
Copyright © 2024, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.
PY - 2024/3/25
Y1 - 2024/3/25
N2 - Decision trees are a fundamental tool in machine learning for representing, classifying, and generalizing data. It is desirable to construct “small” decision trees, by minimizing either the size (s) or the depth (d) of the decision tree (DT). Recently, the parameterized complexity of DECISION TREE LEARNING has attracted a lot of attention. We consider a generalization of DECISION TREE LEARNING where given a classification instance E and an integer t, the task is to find a “small” DT that disagrees with E in at most t examples. We consider two problems: DTSO and DTDO, where the goal is to construct a DT minimizing s and d, respectively. We first establish that both DTSO and DTDO are W[1]-hard when parameterized by s+δmax and d+δmax, respectively, where δmax is the maximum number of features in which two differently labeled examples can differ. We complement this result by showing that these problems become FPT if we include the parameter t. We also consider the kernelization complexity of these problems and establish several positive and negative results for both DTSO and DTDO.
AB - Decision trees are a fundamental tool in machine learning for representing, classifying, and generalizing data. It is desirable to construct “small” decision trees, by minimizing either the size (s) or the depth (d) of the decision tree (DT). Recently, the parameterized complexity of DECISION TREE LEARNING has attracted a lot of attention. We consider a generalization of DECISION TREE LEARNING where given a classification instance E and an integer t, the task is to find a “small” DT that disagrees with E in at most t examples. We consider two problems: DTSO and DTDO, where the goal is to construct a DT minimizing s and d, respectively. We first establish that both DTSO and DTDO are W[1]-hard when parameterized by s+δmax and d+δmax, respectively, where δmax is the maximum number of features in which two differently labeled examples can differ. We complement this result by showing that these problems become FPT if we include the parameter t. We also consider the kernelization complexity of these problems and establish several positive and negative results for both DTSO and DTDO.
UR - http://www.scopus.com/inward/record.url?scp=85189646365&partnerID=8YFLogxK
U2 - 10.1609/aaai.v38i11.29098
DO - 10.1609/aaai.v38i11.29098
M3 - Conference contribution
AN - SCOPUS:85189646365
T3 - Proceedings of the AAAI Conference on Artificial Intelligence
SP - 12100
EP - 12108
BT - Technical Tracks 14
A2 - Wooldridge, Michael
A2 - Dy, Jennifer
A2 - Natarajan, Sriraam
PB - Association for the Advancement of Artificial Intelligence
Y2 - 20 February 2024 through 27 February 2024
ER -