An IRT forecasting model: linking proper scoring rules to item response theory

被引:0
|
作者
Bo, Yuanchao Emily [1 ]
Budescu, David V. [2 ]
Lewis, Charles [2 ]
Tetlock, Philip E. [3 ]
Mellers, Barbara [3 ]
机构
[1] NWEA, Portland, OR 97209 USA
[2] Fordham Univ, Bronx, NY 10458 USA
[3] Univ Penn, Philadelphia, PA 19104 USA
来源
JUDGMENT AND DECISION MAKING | 2017年 / 12卷 / 02期
关键词
IRT; Forecasting; Brier scores; Proper Scoring Rules; Good Judgment Project; Gibbs sampling; DOMINANCE ANALYSIS; PREDICTORS; ACCURACY;
D O I
暂无
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
This article proposes an Item Response Theoretical (IRT) forecasting model that incorporates proper scoring rules and provides evaluations of forecasters' expertise in relation to the features of the specific questions they answer. We illustrate the model using geopolitical forecasts obtained by the Good Judgment Project (GJP) (see Mellers, Ungar, Baron, Ramos, Gurcay, Fincher, Scott, Moore, Atanasov, Swift, Murray, Stone & Tetlock, 2014). The expertise estimates from the IRT model, which take into account variation in the difficulty and discrimination power of the events, capture the underlying construct being measured and are highly correlated with the forecasters' Brier scores. Furthermore, our expertise estimates based on the first three years of the GJP data are better predictors of both the forecasters' fourth year Brier scores and their activity level than the overall Brier scores obtained and Merkle's (2016) predictions, based on the same period. Lastly, we discuss the benefits of using event-characteristic information in forecasting.
引用
收藏
页码:90 / 103
页数:14
相关论文
共 50 条
  • [31] Construct validity: How to prove it by theory-based item generating rules and IRT model-based analyses
    Poinstingl, Herbert
    Kubinger, Klaus D.
    INTERNATIONAL JOURNAL OF PSYCHOLOGY, 2008, 43 (3-4) : 510 - 510
  • [32] IRT ITEM BIAS DETECTION PROCEDURES - ISSUES OF MODEL MISSPECIFICATION, ROBUSTNESS, AND PARAMETER LINKING
    LAUTENSCHLAGER, GJ
    PARK, DG
    APPLIED PSYCHOLOGICAL MEASUREMENT, 1988, 12 (04) : 365 - 376
  • [33] Item response theory linking of the PROMIS and FAAM instruments
    Hung, Man
    Baumhauer, Judith
    Hon, Shirley
    Voss, Maren
    Li, Wei
    Xu, Julie
    Ruiz, Bianca
    Rosales, Megan
    Bounsanga, Jerry
    Su, Weicong
    Saltzman, Charles
    QUALITY OF LIFE RESEARCH, 2018, 27 : S181 - S181
  • [34] Erratum to: Linking Item Response Model Parameters
    Wim J. van der Linden
    Michelle D. Barrett
    Psychometrika, 2017, 82 : 273 - 273
  • [35] IRTEQ: Software for Linking and Equating with Item Response Theory
    Thompson, Nathan A.
    INTERNATIONAL JOURNAL OF TESTING, 2010, 10 (02) : 200 - 206
  • [36] Assessing the Item Response Theory With Covariate (IRT-C) Procedure for Ascertaining Differential Item Functioning
    Tay, Louis
    Vermunt, Jeroen K.
    Wang, Chun
    INTERNATIONAL JOURNAL OF TESTING, 2013, 13 (03) : 201 - 222
  • [37] β3-IRT: A New Item Response Model and its Applications
    Chen, Yu
    Silva Filho, Telmo
    Prudencio, Ricardo B. C.
    Diethe, Tom
    Flach, Peter
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [38] Assessment of diagnostics tools for sarcopenia severity using the item response theory (IRT)
    M. Steffl
    M. Musalek
    V. Kramperova
    M. Petr
    E. Kohlikova
    I. Holmerova
    L. Volicer
    The journal of nutrition, health & aging, 2016, 20 : 1051 - 1055
  • [39] An application of item response time: The effort-moderated IRT model
    Wise, SL
    DeMars, CE
    JOURNAL OF EDUCATIONAL MEASUREMENT, 2006, 43 (01) : 19 - 38
  • [40] Using item response theory (IRT) to develop measures of asthma control.
    Ahmed, S
    Ernst, P
    Colman, N
    Tamblyn, R
    AMERICAN JOURNAL OF EPIDEMIOLOGY, 2005, 161 (11) : S70 - S70