An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization

被引:0
|
作者
Thomas G. Dietterich
机构
[1] Oregon State University,Department of Computer Science
来源
Machine Learning | 2000年 / 40卷
关键词
decision trees; ensemble learning; bagging; boosting; C4.5; Monte Carlo methods;
D O I
暂无
中图分类号
学科分类号
摘要
Bagging and boosting are methods that generate a diverse ensemble of classifiers by manipulating the training data given to a “base” learning algorithm. Breiman has pointed out that they rely for their effectiveness on the instability of the base learning algorithm. An alternative approach to generating an ensemble is to randomize the internal decisions made by the base algorithm. This general approach has been studied previously by Ali and Pazzani and by Dietterich and Kong. This paper compares the effectiveness of randomization, bagging, and boosting for improving the performance of the decision-tree algorithm C4.5. The experiments show that in situations with little or no classification noise, randomization is competitive with (and perhaps slightly superior to) bagging but not as accurate as boosting. In situations with substantial classification noise, bagging is much better than boosting, and sometimes better than randomization.
引用
收藏
页码:139 / 157
页数:18
相关论文
共 50 条
  • [21] Landslide Susceptibility Assessment Using Bagging Ensemble Based Alternating Decision Trees, Logistic Regression and J48 Decision Trees Methods: A Comparative Study
    Pham B.T.
    Tien Bui D.
    Prakash I.
    Geotechnical and Geological Engineering, 2017, 35 (6) : 2597 - 2611
  • [22] Comparison of Decision Tree Classification Methods and Gradient Boosted Trees
    Dikananda, Arif Rinaldi
    Jumini, Sri
    Tarihoran, Nafan
    Christinawati, Santy
    Trimastuti, Wahyu
    Rahim, Robbi
    TEM JOURNAL-TECHNOLOGY EDUCATION MANAGEMENT INFORMATICS, 2022, 11 (01): : 316 - 322
  • [23] Predictive Ensemble Modelling - Experimental Comparison of Boosting Implementation Methods
    Adegoke, Vincent F.
    Chen, Daqing
    Banissi, Safia
    Banissi, Ebad
    UKSIM-AMSS 11TH EUROPEAN MODELLING SYMPOSIUM ON COMPUTER MODELLING AND SIMULATION (EMS 2017), 2017, : 11 - 16
  • [24] Comparison of three methods for estimating internal support on phylogenetic trees
    Mort, ME
    Soltis, PS
    Soltis, DE
    Mabry, ML
    SYSTEMATIC BIOLOGY, 2000, 49 (01) : 160 - 171
  • [25] An Experimental Study on Combining Binarization Techniques and Ensemble Methods of Decision Trees
    Rodriguez, Juan J.
    Diez-Pastor, Jose F.
    Arnaiz-Gonzalez, Alvar
    Garcia-Osorio, Cesar
    MULTIPLE CLASSIFIER SYSTEMS (MCS 2015), 2015, 9132 : 181 - 193
  • [26] On the three levels security policy comparison between SVM and decision trees
    Radi, A.
    Kartit, A.
    Aboutajdine, D.
    Regragui, B.
    El Marraki, M.
    Ramrami, A.
    Journal of Theoretical and Applied Information Technology, 2012, 35 (01) : 56 - 68
  • [27] A comparison of predictive methods in extinction risk studies: Contrasts and decision trees
    Sullivan, Matthew S.
    Jones, Martin J.
    Lee, David C.
    Marsden, Stuart J.
    Fielding, Alan H.
    Young, Emily V.
    BIODIVERSITY AND CONSERVATION, 2006, 15 (06) : 1977 - 1991
  • [28] A Comparison of Predictive Methods in Extinction Risk Studies: Contrasts and Decision Trees
    Matthew S. Sullivan
    Martin J. Jones
    David C. Lee
    Stuart J. Marsden
    Alan H. Fielding
    Emily V. Young
    Biodiversity & Conservation, 2006, 15 : 1977 - 1991
  • [29] An experimental comparison of ensemble learning methods on decision boundaries
    Liu, Y
    Yao, X
    Zhao, QF
    Higuchi, T
    PROCEEDING OF THE 2002 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-3, 2002, : 221 - 226
  • [30] Comparison of laminate stiffness as measured by three experimental methods
    Lasn, Kaspar
    Echtermeyer, Andreas T.
    Klauson, Aleksander
    Chati, Farid
    Decultot, Dominique
    POLYMER TESTING, 2015, 44 : 143 - 152