Maintaining Equivalent Cut Scores for Small Sample Test Forms

被引:8
|
作者
Dwyer, Andrew C. [1 ]
机构
[1] Amer Board Pediat Inc, Psychometr, 111 Silver Cedar Court, Chapel Hill, NC 27514 USA
关键词
D O I
10.1111/jedm.12098
中图分类号
G44 [教育心理学];
学科分类号
0402 ; 040202 ;
摘要
This study examines the effectiveness of three approaches for maintaining equivalent performance standards across test forms with small samples: (1) common-item equating, (2) resetting the standard, and (3) rescaling the standard. Rescaling the standard (i.e., applying common-item equating methodology to standard setting ratings to account for systematic differences between standard setting panels) has received almost no attention in the literature. Identity equating was also examined to provide context. Data from a standard setting form of a large national certification test (N examinees = 4,397; N panelists = 13) were split into content-equivalent subforms with common items, and resampling methodology was used to investigate the error introduced by each approach. Common-item equating (circle-arc and nominal weights mean) was evaluated at samples of size 10, 25, 50, and 100. The standard setting approaches (resetting and rescaling the standard) were evaluated by resampling (N = 8) and by simulating panelists (N = 8, 13, and 20). Results were inconclusive regarding the relative effectiveness of resetting and rescaling the standard. Small-sample equating, however, consistently produced new form cut scores that were less biased and less prone to random error than new form cut scores based on resetting or rescaling the standard.
引用
收藏
页码:3 / 22
页数:20
相关论文
共 50 条