An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization

被引:0
|
作者
Thomas G. Dietterich
机构
[1] Oregon State University,Department of Computer Science
来源
Machine Learning | 2000年 / 40卷
关键词
decision trees; ensemble learning; bagging; boosting; C4.5; Monte Carlo methods;
D O I
暂无
中图分类号
学科分类号
摘要
Bagging and boosting are methods that generate a diverse ensemble of classifiers by manipulating the training data given to a “base” learning algorithm. Breiman has pointed out that they rely for their effectiveness on the instability of the base learning algorithm. An alternative approach to generating an ensemble is to randomize the internal decisions made by the base algorithm. This general approach has been studied previously by Ali and Pazzani and by Dietterich and Kong. This paper compares the effectiveness of randomization, bagging, and boosting for improving the performance of the decision-tree algorithm C4.5. The experiments show that in situations with little or no classification noise, randomization is competitive with (and perhaps slightly superior to) bagging but not as accurate as boosting. In situations with substantial classification noise, bagging is much better than boosting, and sometimes better than randomization.
引用
收藏
页码:139 / 157
页数:18
相关论文
共 50 条
  • [11] An Experimental Study about Simple Decision Trees for Bagging Ensemble on Datasets with Classification Noise
    Abellan, Joaquin
    Masegosa, Andres R.
    SYMBOLIC AND QUANTITATIVE APPROACHES TO REASONING WITH UNCERTAINTY, PROCEEDINGS, 2009, 5590 : 446 - 456
  • [12] Comparison of Bagging and Boosting Ensemble Machine Learning Methods for Automated EMG Signal Classification
    Yaman, Emine
    Subasi, Abdulhamit
    BIOMED RESEARCH INTERNATIONAL, 2019, 2019
  • [13] A comparison of three methods of constructing Lyapunov functions
    Rumyantsev, VV
    PMM JOURNAL OF APPLIED MATHEMATICS AND MECHANICS, 1995, 59 (06): : 873 - 877
  • [14] A Review and Experimental Comparison of Multivariate Decision Trees
    Canete-Sifuentes, Leonardo
    Monroy, Raul
    Medina-Perez, Miguel Angel
    IEEE ACCESS, 2021, 9 : 110451 - 110479
  • [15] An Application of Hybrid Bagging-Boosting Decision Trees Ensemble Model for Riverine Flood Susceptibility Mapping and Regional Risk DelineationAn Application of Hybrid Bagging-Boosting Decision Trees Ensemble Model for Riverine Flood Susceptibility Mapping and Regional Risk Delineation
    Javeria Sarwar
    Saud Ahmed Khan
    Muhammad Azmat
    Faridoon Khan
    Water Resources Management, 2025, 39 (2) : 547 - 577
  • [16] A Comparison of Three Voting Methods for Bagging with the MLEM2 Algorithm
    Cohagan, Clinton
    Grzymala-Busse, Jerzy W.
    Hippe, Zdzislaw S.
    INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING - IDEAL 2010, 2010, 6283 : 118 - +
  • [17] Comparison of Three Methods for Constructing Real Driving Cycles
    Ignacio Huertas, Jose I.
    Felipe Quirama, Luis
    Giraldo, Michael
    Diaz, Jenny
    ENERGIES, 2019, 12 (04)
  • [18] A comparison of three approaches for constructing robust experimental designs
    Agboto V.
    Li W.
    Nachtsheim C.
    Journal of Statistical Theory and Practice, 2011, 5 (1) : 1 - 11
  • [19] An Application of Hybrid Bagging-Boosting Decision Trees Ensemble Model for Riverine Flood Susceptibility Mapping and Regional Risk Delineation
    Sarwar, Javeria
    Khan, Saud Ahmed
    Azmat, Muhammad
    Khan, Faridoon
    WATER RESOURCES MANAGEMENT, 2025, 39 (02) : 547 - 577
  • [20] Experimental comparison of classification uncertainty for randomised and Bayesian Decision Tree ensembles
    Schetinin, V
    Partridge, D
    Krzanowski, WJ
    Everson, RM
    Fieldsend, JE
    Bailey, TC
    Hernandez, A
    INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING IDEAL 2004, PROCEEDINGS, 2004, 3177 : 726 - 732