TY - JOUR
T1 - Keeping up with the times
T2 - Revising and refreshing a rating scale
AU - Banerjee, Jayanti
AU - Yan, Xun
AU - Chapman, Mark
AU - Elliott, Heather
N1 - Publisher Copyright:
© 2015 Elsevier Inc.
PY - 2015/10/1
Y1 - 2015/10/1
N2 - In performance-based writing assessment, regular monitoring and modification of the rating scale is essential to ensure reliable test scores and valid score inferences. However, the development and modification of rating scales (particularly writing scales) is rarely discussed in language assessment literature. The few studies documenting the scale development process have derived the rating scale from analyzing one or two data sources: expert intuition, rater discussion, and/or real performance. This study reports on the review and revision of a rating scale for the writing section of a large-scale, advanced-level English language proficiency examination. Specifically, this study first identified from literature, the features of written text that tend to reliably distinguish between essays across levels of proficiency. Next, using corpus-based tools, 796 essays were analyzed for text features that predict writing proficiency levels. Lastly, rater discussions were analyzed to identify components of the existing scale that raters found helpful for assigning scores. Based on these findings, a new rating scale has been prepared. The results of this work demonstrate the benefits of triangulating information from writing research, rater discussions, and real performances in rating scale design.
AB - In performance-based writing assessment, regular monitoring and modification of the rating scale is essential to ensure reliable test scores and valid score inferences. However, the development and modification of rating scales (particularly writing scales) is rarely discussed in language assessment literature. The few studies documenting the scale development process have derived the rating scale from analyzing one or two data sources: expert intuition, rater discussion, and/or real performance. This study reports on the review and revision of a rating scale for the writing section of a large-scale, advanced-level English language proficiency examination. Specifically, this study first identified from literature, the features of written text that tend to reliably distinguish between essays across levels of proficiency. Next, using corpus-based tools, 796 essays were analyzed for text features that predict writing proficiency levels. Lastly, rater discussions were analyzed to identify components of the existing scale that raters found helpful for assigning scores. Based on these findings, a new rating scale has been prepared. The results of this work demonstrate the benefits of triangulating information from writing research, rater discussions, and real performances in rating scale design.
KW - Corpora
KW - Discriminant function analysis
KW - Rating scale design
KW - Scale validation
UR - http://www.scopus.com/inward/record.url?scp=84941992668&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84941992668&partnerID=8YFLogxK
U2 - 10.1016/j.asw.2015.07.001
DO - 10.1016/j.asw.2015.07.001
M3 - Article
AN - SCOPUS:84941992668
SN - 1075-2935
VL - 26
SP - 5
EP - 19
JO - Assessing Writing
JF - Assessing Writing
ER -