A Unified Framework for Decision Tree on Continuous Attributes

被引:5
|
作者
Yan, Jianjian [1 ]
Zhang, Zhongnan [1 ]
Xie, Lingwei [1 ]
Zhu, Zhantu [1 ]
机构
[1] Xiamen Univ, Software Sch, Xiamen 361005, Peoples R China
来源
IEEE ACCESS | 2019年 / 7卷
关键词
Decision tree; classification; unified framework; split criteria; CLASSIFICATION; NETWORKS; SVM;
D O I
10.1109/ACCESS.2019.2892083
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The standard algorithms of decision trees and their derived methods are usually constructed on the basis of the frequency information. However, they still suffer from a dilemma or multichotomous question for continuous attributes when two or more candidate cut points have the same or similar splitting performance with the optimal value, such as the maximal information gain ratio or the minimal Gini index. In this paper, we propose a unified framework model to deal with this question. We then design two algorithms based on Splitting Performance and the number of Expected Segments, called SPES1 and SPES2, which determine the optimal cut point, as follows. First, several candidate cut points are selected based on their splitting performances being the closest to the optimal. Second, we compute the number of expected segments for each candidate cut point. Finally, we combine these two measures by introducing a weighting factor alpha to determine the optimal one from several candidate cut points. To validate the effectiveness of our methods, we perform them on 25 benchmark datasets. The experimental results demonstrate that the classification accuracies of the proposed algorithms are superior to the current state-of-the-art methods in tackling the multichotomous question, about 5% in some cases. In particular, according to the proposed methods, the number of candidate cut points converges to a certain extent.
引用
收藏
页码:11924 / 11933
页数:10
相关论文
共 50 条
  • [1] Constructing decision tree with continuous attributes for binary classification
    Jiang, YH
    Zhou, RF
    Yang, XJ
    2002 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-4, PROCEEDINGS, 2002, : 617 - 622
  • [2] Numerical Continuous-valued Attributes Semanteme in Decision Tree
    Shi, Nianyun
    Chu, Xi
    Zhou, Xiaotian
    ICNC 2008: FOURTH INTERNATIONAL CONFERENCE ON NATURAL COMPUTATION, VOL 2, PROCEEDINGS, 2008, : 642 - +
  • [3] Segment Based Decision Tree Induction With Continuous Valued Attributes
    Wang, Ran
    Kwong, Sam
    Wang, Xi-Zhao
    Jiang, Qingshan
    IEEE TRANSACTIONS ON CYBERNETICS, 2015, 45 (07) : 1262 - 1275
  • [4] Side Effect of Cut in Decision Tree Generation for Continuous Attributes
    Wang, Xi-Zhao
    Gao, Xiang-Hui
    He, Qiang
    IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC 2010), 2010,
  • [5] ON THE HANDLING OF CONTINUOUS-VALUED ATTRIBUTES IN DECISION TREE GENERATION
    FAYYAD, UM
    IRANI, KB
    MACHINE LEARNING, 1992, 8 (01) : 87 - 102
  • [6] On a unified framework for sampling with and without replacement in decision tree ensembles
    Martinez-Otzeta, J. M.
    Sierra, B.
    Lazkano, E.
    Jauregi, E.
    ARTIFICIAL INTELLIGENCE: METHODOLOGY, SYSTEMS, AND APPLICATIONS, PROCEEDINGS, 2006, 4183 : 118 - 127
  • [7] On the handling of fuzziness for continuous-valued attributes in decision tree generation
    Wang, XZ
    Hong, JR
    FUZZY SETS AND SYSTEMS, 1998, 99 (03) : 283 - 290
  • [8] A note on the handling of fuzziness for continuous-valued attributes in decision tree generation
    Hong, Dug Hun
    Lee, Sungho
    Kim, Kyung Tae
    FUZZY SYSTEMS AND KNOWLEDGE DISCOVERY, PROCEEDINGS, 2006, 4223 : 241 - 245
  • [9] MC4.5 decision tree algorithm: An improved use of continuous attributes
    Cherfi, Anis
    Nouira, Kaouther
    Ferchichi, Ahmed
    International Journal of Computational Intelligence Studies, 2020, 9 (1-2) : 4 - 17
  • [10] Handling continuous-valued attributes in decision tree with neural network modeling
    Kim, D
    Lee, J
    MACHINE LEARNING: ECML 2000, 2000, 1810 : 211 - 219