An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization

被引:0
|
作者
Thomas G. Dietterich
机构
[1] Oregon State University,Department of Computer Science
来源
Machine Learning | 2000年 / 40卷
关键词
decision trees; ensemble learning; bagging; boosting; C4.5; Monte Carlo methods;
D O I
暂无
中图分类号
学科分类号
摘要
Bagging and boosting are methods that generate a diverse ensemble of classifiers by manipulating the training data given to a “base” learning algorithm. Breiman has pointed out that they rely for their effectiveness on the instability of the base learning algorithm. An alternative approach to generating an ensemble is to randomize the internal decisions made by the base algorithm. This general approach has been studied previously by Ali and Pazzani and by Dietterich and Kong. This paper compares the effectiveness of randomization, bagging, and boosting for improving the performance of the decision-tree algorithm C4.5. The experiments show that in situations with little or no classification noise, randomization is competitive with (and perhaps slightly superior to) bagging but not as accurate as boosting. In situations with substantial classification noise, bagging is much better than boosting, and sometimes better than randomization.
引用
收藏
页码:139 / 157
页数:18
相关论文
共 50 条
  • [1] An experimental comparison of three methods for constructing ensembles of decision trees: Bagging, boosting, and randomization
    Dietterich, TG
    MACHINE LEARNING, 2000, 40 (02) : 139 - 157
  • [2] A comparison of stacking with meta decision trees to bagging, boosting, and stacking with other methods
    Zenko, B
    Todorovski, L
    Dzeroski, S
    2001 IEEE INTERNATIONAL CONFERENCE ON DATA MINING, PROCEEDINGS, 2001, : 669 - 670
  • [3] Comparing Boosting and Bagging for Decision Trees of Rankings
    Antonella Plaia
    Simona Buscemi
    Johannes Fürnkranz
    Eneldo Loza Mencía
    Journal of Classification, 2022, 39 : 78 - 99
  • [4] Comparing Boosting and Bagging for Decision Trees of Rankings
    Plaia, Antonella
    Buscemi, Simona
    Fuernkranz, Johannes
    Mencia, Eneldo Loza
    JOURNAL OF CLASSIFICATION, 2022, 39 (01) : 78 - 99
  • [5] Comparison of Bagging, Boosting and Stacking Ensembles Applied to Real Estate Appraisal
    Graczyk, Magdalena
    Lasota, Tadeusz
    Trawinski, Bogdan
    Trawinski, Krzysztof
    INTELLIGENT INFORMATION AND DATABASE SYSTEMS, PT II, PROCEEDINGS, 2010, 5991 : 340 - +
  • [6] Constructing ensembles of ASR systems using randomized decision trees
    Siohan, O
    Ramabhadran, B
    Kingsbury, B
    2005 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS 1-5: SPEECH PROCESSING, 2005, : 197 - 200
  • [7] Improvement of Best First Decision Trees Using Bagging and Dagging Ensembles for Flood Probability Mapping
    Yariyan, Peyman
    Janizadeh, Saeid
    Phong Van Tran
    Huu Duy Nguyen
    Costache, Romulus
    Hiep Van Le
    Binh Thai Pham
    Pradhan, Biswajeet
    Tiefenbacher, John P.
    WATER RESOURCES MANAGEMENT, 2020, 34 (09) : 3037 - 3053
  • [8] Improvement of Best First Decision Trees Using Bagging and Dagging Ensembles for Flood Probability Mapping
    Peyman Yariyan
    Saeid Janizadeh
    Tran Van Phong
    Huu Duy Nguyen
    Romulus Costache
    Hiep Van Le
    Binh Thai Pham
    Biswajeet Pradhan
    John P. Tiefenbacher
    Water Resources Management, 2020, 34 : 3037 - 3053
  • [9] An experimental comparison of Random Projection ensembles with linear kernel SVMs and Bagging and BagBoosting methods for the classification of gene expression data
    Folgieri, Raffaella
    NEURAL NETS WIRN09, 2009, 204 : 208 - 216
  • [10] Histogram-Based Algorithm for Building Gradient Boosting Ensembles of Piecewise Linear Decision Trees
    Guryanov, Aleksei
    ANALYSIS OF IMAGES, SOCIAL NETWORKS AND TEXTS, AIST 2019, 2019, 11832 : 39 - 50