TY - JOUR
T1 - Exploring collaborative caption editing to augment video-based learning
AU - Bhavya, Bhavya
AU - Chen, Si
AU - Zhang, Zhilin
AU - Li, Wenting
AU - Zhai, Chengxiang
AU - Angrave, Lawrence
AU - Huang, Yun
N1 - Publisher Copyright:
© 2022, Association for Educational Communications and Technology.
PY - 2022/10
Y1 - 2022/10
N2 - Captions play a major role in making educational videos accessible to all and are known to benefit a wide range of learners. However, many educational videos either do not have captions or have inaccurate captions. Prior work has shown the benefits of using crowdsourcing to obtain accurate captions in a cost-efficient way, though there is a lack of understanding of how learners edit captions of educational videos either individually or collaboratively. In this work, we conducted a user study where 58 learners (in a course of 387 learners) participated in the editing of captions in 89 lecture videos that were generated by Automatic Speech Recognition (ASR) technologies. For each video, different learners conducted two rounds of editing. Based on editing logs, we created a taxonomy of errors in educational video captions (e.g., Discipline-Specific, General, Equations). From the interviews, we identified individual and collaborative error editing strategies. We then further demonstrated the feasibility of applying machine learning models to assist learners in editing. Our work provides practical implications for advancing video-based learning and for educational video caption editing.
AB - Captions play a major role in making educational videos accessible to all and are known to benefit a wide range of learners. However, many educational videos either do not have captions or have inaccurate captions. Prior work has shown the benefits of using crowdsourcing to obtain accurate captions in a cost-efficient way, though there is a lack of understanding of how learners edit captions of educational videos either individually or collaboratively. In this work, we conducted a user study where 58 learners (in a course of 387 learners) participated in the editing of captions in 89 lecture videos that were generated by Automatic Speech Recognition (ASR) technologies. For each video, different learners conducted two rounds of editing. Based on editing logs, we created a taxonomy of errors in educational video captions (e.g., Discipline-Specific, General, Equations). From the interviews, we identified individual and collaborative error editing strategies. We then further demonstrated the feasibility of applying machine learning models to assist learners in editing. Our work provides practical implications for advancing video-based learning and for educational video caption editing.
KW - Caption transcription
KW - Collaborative editing
KW - Lecture video caption editing
KW - Technology-assisted editing
UR - http://www.scopus.com/inward/record.url?scp=85134311910&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85134311910&partnerID=8YFLogxK
U2 - 10.1007/s11423-022-10137-5
DO - 10.1007/s11423-022-10137-5
M3 - Article
C2 - 35855355
AN - SCOPUS:85134311910
SN - 1042-1629
VL - 70
SP - 1755
EP - 1779
JO - Educational Technology Research and Development
JF - Educational Technology Research and Development
IS - 5
ER -