Rater characteristics, response content, and scoring contexts: Decomposing the determinates of scoring accuracy

Frontiers in Psychology 13 (2022)
  Copy   BIBTEX

Abstract

Raters may introduce construct-irrelevant variance when evaluating written responses to performance assessments, threatening the validity of students’ scores. Numerous factors in the rating process, including the content of students’ responses, the characteristics of raters, and the context in which the scoring occurs, are thought to influence the quality of raters’ scores. Despite considerable study of rater effects, little research has examined the relative impacts of the factors that influence rater accuracy. In practice, such integrated examinations are needed to afford evidence-based decisions of rater selection, training, and feedback. This study provides the first naturalistic, integrated examination of rater accuracy in a large-scale assessment program. Leveraging rater monitoring data from an English language arts summative assessment program, I specified cross-classified, multilevel models via Bayesian estimation to decompose the impact of response content, rater characteristics, and scoring contexts on rater accuracy. Results showed relatively little variation in accuracy attributable to teams, items, and raters. Raters did not collectively exhibit differential accuracy over time, though there was significant variation in individual rater’s scoring accuracy from response to response and day to day. I found considerable variation in accuracy across responses, which was in part explained by text features and other measures of response content that influenced scoring difficulty. Some text features differentially influenced the difficulty of scoring research and writing content. Multiple measures of raters’ qualification performance predicted their scoring accuracy, but general rater background characteristics including experience and education did not. Site-based and remote raters demonstrated comparable accuracy, while evening-shift raters were slightly less accurate, on average, than day-shift raters. This naturalistic, integrated examination of rater accuracy extends previous research and provides implications for rater recruitment, training, monitoring, and feedback to improve human evaluation of written responses.

Links

PhilArchive



    Upload a copy of this work     Papers currently archived: 93,867

External links

Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Similar books and articles

Accuracy, Verisimilitude, and Scoring Rules.Jeffrey Dunn - 2019 - Australasian Journal of Philosophy 97 (1):151-166.
Refinement: Measuring informativeness of ratings in the absence of a gold standard.Sheridan Grant, Marina Meilă, Elena Erosheva & Carole Lee - 2022 - British Journal of Mathematical and Statistical Psychology 75 (3):593-615.

Analytics

Added to PP
2022-08-11

Downloads
6 (#1,479,581)

6 months
4 (#1,004,663)

Historical graph of downloads
How can I increase my downloads?