TY - GEN
T1 - What's the Point? How Scores Undermine Written Comments on Open-Ended Work
AU - Crain, Patrick
AU - Bailey, Brian
N1 - Publisher Copyright:
© 2021 ACM.
PY - 2021/6/8
Y1 - 2021/6/8
N2 - Scaling assessments typically relies on quantifying work quality, yet written comments are the assessment method of choice for open-ended work. Existing scalable solutions compromise by mapping quality scores to pre-authored comments, but how scores influence interpretation of these comments is not well understood. We report results from a study of how 441 participants authored and revised short stories in response to a score, written comments, both types of feedback, or no feedback. We analyzed data from the story-writing task and two surveys to determine task and feedback satisfaction, revision depth and effort, and improvement between drafts for each participant. We found task satisfaction and task performance were positively correlated among participants who were shown a score. Feedback satisfaction, revision effort, and improvement were highest among participants shown written comments. Either type of feedback prompted more deep revisions than no feedback, but together elicited fewer deep revisions than written comments alone. Our work informs the design of scalable open-ended assessment systems by contributing insights regarding how scores influence perceptions of written feedback and subsequent revision outcomes.
AB - Scaling assessments typically relies on quantifying work quality, yet written comments are the assessment method of choice for open-ended work. Existing scalable solutions compromise by mapping quality scores to pre-authored comments, but how scores influence interpretation of these comments is not well understood. We report results from a study of how 441 participants authored and revised short stories in response to a score, written comments, both types of feedback, or no feedback. We analyzed data from the story-writing task and two surveys to determine task and feedback satisfaction, revision depth and effort, and improvement between drafts for each participant. We found task satisfaction and task performance were positively correlated among participants who were shown a score. Feedback satisfaction, revision effort, and improvement were highest among participants shown written comments. Either type of feedback prompted more deep revisions than no feedback, but together elicited fewer deep revisions than written comments alone. Our work informs the design of scalable open-ended assessment systems by contributing insights regarding how scores influence perceptions of written feedback and subsequent revision outcomes.
KW - automated assessment
KW - creative work
KW - open-ended work
KW - scaling assessment
UR - http://www.scopus.com/inward/record.url?scp=85108080730&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85108080730&partnerID=8YFLogxK
U2 - 10.1145/3430895.3460132
DO - 10.1145/3430895.3460132
M3 - Conference contribution
AN - SCOPUS:85108080730
T3 - L@S 2021 - Proceedings of the 8th ACM Conference on Learning @ Scale
SP - 127
EP - 138
BT - L@S 2021 - Proceedings of the 8th ACM Conference on Learning @ Scale
PB - Association for Computing Machinery
T2 - 8th Annual ACM Conference on Learning at Scale, L@S 2021
Y2 - 22 June 2021 through 25 June 2021
ER -