Improving over-fitting in ensemble regression by imprecise probabilities

被引:19
|
作者
Utkin, Lev V. [1 ]
Wiencierz, Andrea [2 ]
机构
[1] St Petersburg State Forest Tech Univ, Dept Control Automat & Syst Anal, St Petersburg 194021, Russia
[2] Univ York, Dept Math, York YO10 5DD, N Yorkshire, England
关键词
Regression; AdaBoost algorithm; Over-fitting; Linear-vacuous mixture model; Kolmogorov-Smirnov bounds; ADABOOST.RT; PERFORMANCE;
D O I
10.1016/j.ins.2015.04.037
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, generalized versions of two ensemble methods for regression based on variants of the original AdaBoost algorithm are proposed. The generalization of these regression methods consists in restricting the unit simplex for the weights of the instances to a smaller set of weighting probabilities. Various imprecise statistical models can be used to obtain a restricted set of weighting probabilities, whose sizes each depend on a single parameter. For particular choices of this parameter, the proposed algorithms reduce to standard AdaBoost-based regression algorithms or to standard regression. The main advantage of the proposed algorithms compared to the basic AdaBoost-based regression methods is that they have less tendency to over-fitting, because the weights of the hard instances are restricted. Several simulations and applications furthermore indicate a better performance of the proposed regression methods in comparison with the corresponding standard regression methods. (C) 2015 Elsevier Inc. All rights reserved.
引用
收藏
页码:315 / 328
页数:14
相关论文
共 50 条