The Effect of Numbers of Experts and Common Items on Cutting Score Equivalents Based on Expert Judgment
Open Access
- 1 September 1991
- journal article
- research article
- Published by SAGE Publications in Applied Psychological Measurement
- Vol. 15 (3), 241-246
- https://doi.org/10.1177/014662169101500303
Abstract
The effect of different numbers of experts and common items on the scaling of cutting scores derived by experts' judgments was investigated. Four test forms were created from each of two examinations; each form from the first examina tion shared a block of items with one form from the second examination. Small groups of experts set standards on each using a modification of Angoff's (1971) method. Cutting score equivalents were estimated for the matched forms using dif ferent group sizes and numbers of common items; they were compared with cutting score equivalents based on score equating. Results showed that a reduction in error is associated with using more experts or having more items in common between the two forms. For 25 or more common items and five or more judges, the error was about one item on a 100-item test. More than five experts or 25 common items made only a very small difference in error.Keywords
This publication has 5 references indexed in Scilit:
- Equivalent Pass/Fail DecisionsJournal of Educational Measurement, 1990
- A Note on the Application of Multiple Matrix Sampling to Standard SettingJournal of Educational Measurement, 1988
- A Comparison of Three Variations on a Standard‐Setting MethodJournal of Educational Measurement, 1987
- IRT Test Equating: Relevant Issues and a Review of Recent ResearchReview of Educational Research, 1986
- Theory of mental tests.Published by American Psychological Association (APA) ,1950