TY - CHAP
T1 - Multi-pruning and restricted multi-pruning of decision trees
AU - Alsolami, Fawaz
AU - Azad, Mohammad
AU - Chikalov, Igor
AU - Moshkov, Mikhail
N1 - Publisher Copyright:
© 2020, Springer Nature Switzerland AG.
PY - 2020
Y1 - 2020
N2 - In this chapter, we consider two questions related to decision trees: (i) how to construct decision trees with reasonable number of nodes and reasonable number of misclassifications when they are used for knowledge representation, and (ii) how to improve the prediction accuracy of decision trees when they are used as classifiers. We created so-called multi-pruning approach based on dynamic programming algorithms for bi-criteria optimization of CART-like decision trees relative to the number of nodes and the number of misclassifications. This approach allows us to construct the set of all Pareto optimal points and to derive, for each such point, decision trees with parameters corresponding to that point. Experiments with decision tables from the UCI ML Repository show that, very often, we can find a suitable Pareto optimal point and derive a decision tree with small number of nodes at the expense of small increment in the number of misclassifications. Multi-pruning approach includes a procedure which constructs decision trees that, as classifiers, often outperform decision trees constructed by CART. We considered a modification of multi-pruning approach (restricted multi-pruning) that requires less memory and time but usually keeps the quality of constructed trees as classifiers or as a way for knowledge representation. Based on the uncertainty measure abs which is applicable both to decision tables with single- and many-valued decisions, we extended the considered approaches to the case of decision tables with many-valued decisions.
AB - In this chapter, we consider two questions related to decision trees: (i) how to construct decision trees with reasonable number of nodes and reasonable number of misclassifications when they are used for knowledge representation, and (ii) how to improve the prediction accuracy of decision trees when they are used as classifiers. We created so-called multi-pruning approach based on dynamic programming algorithms for bi-criteria optimization of CART-like decision trees relative to the number of nodes and the number of misclassifications. This approach allows us to construct the set of all Pareto optimal points and to derive, for each such point, decision trees with parameters corresponding to that point. Experiments with decision tables from the UCI ML Repository show that, very often, we can find a suitable Pareto optimal point and derive a decision tree with small number of nodes at the expense of small increment in the number of misclassifications. Multi-pruning approach includes a procedure which constructs decision trees that, as classifiers, often outperform decision trees constructed by CART. We considered a modification of multi-pruning approach (restricted multi-pruning) that requires less memory and time but usually keeps the quality of constructed trees as classifiers or as a way for knowledge representation. Based on the uncertainty measure abs which is applicable both to decision tables with single- and many-valued decisions, we extended the considered approaches to the case of decision tables with many-valued decisions.
UR - http://www.scopus.com/inward/record.url?scp=85063753534&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-12854-8_10
DO - 10.1007/978-3-030-12854-8_10
M3 - Chapter
AN - SCOPUS:85063753534
T3 - Intelligent Systems Reference Library
SP - 153
EP - 174
BT - Intelligent Systems Reference Library
PB - Springer Science and Business Media Deutschland GmbH
ER -