The role of hyperparameters in machine learning models and how to tune them

被引:14
|
作者
Arnold, Christian [1 ]
Biedebach, Luka [2 ]
Kuepfer, Andreas [3 ]
Neunhoeffer, Marcel [4 ,5 ]
机构
[1] Cardiff Univ, Dept Polit & Int Relat, Cardiff, Wales
[2] Reykjavik Univ, Dept Comp Sci, Reykjavik, Iceland
[3] Tech Univ Darmstadt, Inst Polit Sci, Darmstadt, Germany
[4] Boston Univ, Rafik B Hariri Inst Comp & Computat Sci & Engn, Boston, MA 02215 USA
[5] Ludwig Maximilians Univ Munchen, Dept Stat, Munich, Germany
关键词
Best Practice; Hyperparameter Optimization; Machine Learning; REPLICATION;
D O I
10.1017/psrm.2023.61
中图分类号
D0 [政治学、政治理论];
学科分类号
0302 ; 030201 ;
摘要
Hyperparameters critically influence how well machine learning models perform on unseen, out-of-sample data. Systematically comparing the performance of different hyperparameter settings will often go a long way in building confidence about a model's performance. However, analyzing 64 machine learning related manuscripts published in three leading political science journals (APSR, PA, and PSRM) between 2016 and 2021, we find that only 13 publications (20.31 percent) report the hyperparameters and also how they tuned them in either the paper or the appendix. We illustrate the dangers of cursory attention to model and tuning transparency in comparing machine learning models' capability to predict electoral violence from tweets. The tuning of hyperparameters and their documentation should become a standard component of robustness checks for machine learning models.
引用
收藏
页码:841 / 848
页数:8
相关论文
共 50 条