Posterior predictive assessment of item response theory models

被引:141
|
作者
Sinharay, Sandip
Johnson, Matthew S.
Stern, Hal S.
机构
[1] Educ Testing Serv, Princeton, NJ 08541 USA
[2] CUNY Bernard M Baruch Coll, New York, NY 10010 USA
[3] Univ Calif Irvine, Irvine, CA 92717 USA
关键词
Bayesian methods; discrepancy measures; model checking; odds ratio; p values;
D O I
10.1177/0146621605285517
中图分类号
O1 [数学]; C [社会科学总论];
学科分类号
03 ; 0303 ; 0701 ; 070101 ;
摘要
Model checking in item response theory (IRT) is an underdeveloped area. There is no universally accepted tool for checking IRT models. The posterior predictive model-checking method is a popular Bayesian model-checking tool because it has intuitive appeal, is simple to apply, has a strong theoretical basis, and can provide graphical or numerical evidence about model misfit. An important issue with the application of the posterior predictive model-checking method is the choice of a discrepancy measure (which plays a role like that of a test statistic in traditional hypothesis tests). This article examines the performance of a number of discrepancy measures for assessing different aspects of fit of the common IRT models and makes specific recommendations about what measures are most useful in assessing model fit. Graphical summaries of model-checking results are demonstrated to provide useful insights about model fit.
引用
收藏
页码:298 / 321
页数:24
相关论文
共 50 条