کد مقاله | کد نشریه | سال انتشار | مقاله انگلیسی | نسخه تمام متن |
---|---|---|---|---|
344246 | 617358 | 2015 | 15 صفحه PDF | دانلود رایگان |
• We review and revise a writing rating scale for an advanced-level proficiency exam.
• We triangulate findings from research, rater discussions, and corpus analysis.
• We provide theory- and data-driven justifications for revising rating scale criteria.
• The outcomes are directly relevant to the construct definition.
• The new scale better supports the rating process and yields diagnostic information.
In performance-based writing assessment, regular monitoring and modification of the rating scale is essential to ensure reliable test scores and valid score inferences. However, the development and modification of rating scales (particularly writing scales) is rarely discussed in language assessment literature. The few studies documenting the scale development process have derived the rating scale from analyzing one or two data sources: expert intuition, rater discussion, and/or real performance.This study reports on the review and revision of a rating scale for the writing section of a large-scale, advanced-level English language proficiency examination. Specifically, this study first identified from literature, the features of written text that tend to reliably distinguish between essays across levels of proficiency. Next, using corpus-based tools, 796 essays were analyzed for text features that predict writing proficiency levels. Lastly, rater discussions were analyzed to identify components of the existing scale that raters found helpful for assigning scores. Based on these findings, a new rating scale has been prepared. The results of this work demonstrate the benefits of triangulating information from writing research, rater discussions, and real performances in rating scale design.
Journal: Assessing Writing - Volume 26, October 2015, Pages 5–19