On the Value of Ensemble Effort Estimation

被引:159
|
作者
Kocaguneli, Ekrem [1 ]
Menzies, Tim [1 ]
Keung, Jacky W. [2 ]
机构
[1] W Virginia Univ, Dept Comp Sci & Elect Engn, Morgantown, WV 26506 USA
[2] Hong Kong Polytech Univ, Dept Comp, Kowloon, Hong Kong, Peoples R China
基金
美国国家科学基金会;
关键词
Software cost estimation; ensemble; machine learning; regression trees; support vector machines; neural nets; analogy; k-NN; COST ESTIMATION; EMPIRICAL VALIDATION; SOFTWARE; PREDICTION; SELECTION; MODEL; ALGORITHMS; REGRESSION; SYSTEMS;
D O I
10.1109/TSE.2011.111
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Background: Despite decades of research, there is no consensus on which software effort estimation methods produce the most accurate models. Aim: Prior work has reported that, given M estimation methods, no single method consistently outperforms all others. Perhaps rather than recommending one estimation method as best, it is wiser to generate estimates from ensembles of multiple estimation methods. Method: Nine learners were combined with 10 preprocessing options to generate 9 x 10 90 solo methods. These were applied to 20 datasets and evaluated using seven error measures. This identified the best n (in our case n 13) solo methods that showed stable performance across multiple datasets and error measures. The top 2, 4, 8, and 13 solo methods were then combined to generate 12 multimethods, which were then compared to the solo methods. Results: 1) The top 10 (out of 12) multimethods significantly outperformed all 90 solo methods. 2) The error rates of the multimethods were significantly less than the solo methods. 3) The ranking of the best multimethod was remarkably stable. Conclusion: While there is no best single effort estimation method, there exist best combinations of such effort estimation methods.
引用
收藏
页码:1403 / 1416
页数:14
相关论文
共 50 条
  • [1] On the value of parameter tuning in stacking ensemble model for software regression test effort estimation
    Taher Labidi
    Zaineb Sakhrawi
    [J]. The Journal of Supercomputing, 2023, 79 : 17123 - 17145
  • [2] On the value of parameter tuning in stacking ensemble model for software regression test effort estimation
    Labidi, Taher
    Sakhrawi, Zaineb
    [J]. JOURNAL OF SUPERCOMPUTING, 2023, 79 (15): : 17123 - 17145
  • [3] Ensemble effort estimation for novice agile teams
    Alsaadi, Bashaer
    Saeedi, Kawther
    [J]. INFORMATION AND SOFTWARE TECHNOLOGY, 2024, 170
  • [4] Evaluating ensemble imputation in software effort estimation
    Ibtissam Abnane
    Ali Idri
    Imane Chlioui
    Alain Abran
    [J]. Empirical Software Engineering, 2023, 28
  • [5] Evaluating ensemble imputation in software effort estimation
    Abnane, Ibtissam
    Idri, Ali
    Chlioui, Imane
    Abran, Alain
    [J]. EMPIRICAL SOFTWARE ENGINEERING, 2023, 28 (02)
  • [6] Systematic Mapping Study of Ensemble Effort Estimation
    Idri, Ali
    Hosni, Mohamed
    Abran, Alain
    [J]. ENASE: PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON EVALUATION OF NOVEL SOFTWARE APPROACHES TO SOFTWARE ENGINEERING, 2016, : 132 - 139
  • [7] Systematic literature review of ensemble effort estimation
    Idri, Ali
    Hosni, Mohamed
    Abran, Alain
    [J]. JOURNAL OF SYSTEMS AND SOFTWARE, 2016, 118 : 151 - 175
  • [8] Ensemble Effort Estimation using dynamic selection
    Cabral, Jose Thiago H. de A.
    Oliveira, Adriano L. I.
    [J]. JOURNAL OF SYSTEMS AND SOFTWARE, 2021, 175
  • [9] Ensemble effort estimation using selection and genetic algorithms
    Jodpimai, Pichai
    Sophatsathit, Peraphon
    Lursinsap, Chidchanok
    [J]. INTERNATIONAL JOURNAL OF COMPUTER APPLICATIONS IN TECHNOLOGY, 2018, 58 (01) : 17 - 28
  • [10] Heterogeneous Ensemble Imputation for Software Development Effort Estimation
    Abnane, Ibtissam
    Idri, Ali
    Hosni, Mohamed
    Abran, Alain
    [J]. PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON PREDICTIVE MODELS AND DATA ANALYTICS IN SOFTWARE ENGINEERING (PROMISE '21), 2021, : 1 - 10