TY - GEN
T1 - Detecting Programming Plans in Open-ended Code Submissions
AU - Demirtaş, Mehmet Arif
AU - Zheng, Claire
AU - Cunningham, Kathryn
N1 - Publisher Copyright:
© 2025 Copyright held by the owner/author(s).
PY - 2025/2/18
Y1 - 2025/2/18
N2 - Open-ended code-writing exercises are commonly used in large-scale introductory programming courses, as they can be autograded against test cases. However, code writing requires many skills at once, from planning out a solution to applying the intricacies of syntax. As autograding only evaluates code correctness, feedback addressing each of these skills separately cannot be provided. In this work, we explore methods to detect which high-level patterns (i.e. programming plans) have been used in a submission, so learners can receive feedback on planning skills even when their code is not completely correct. Our preliminary results show that LLMs with few-shot prompting can detect the use of programming plans in 95% of correct and 86% of partially correct submissions. Incorporating LLMs into grading of open-ended programming exercises can enable more fine-grained feedback to students, even in cases where their code does not compile due to other errors.
AB - Open-ended code-writing exercises are commonly used in large-scale introductory programming courses, as they can be autograded against test cases. However, code writing requires many skills at once, from planning out a solution to applying the intricacies of syntax. As autograding only evaluates code correctness, feedback addressing each of these skills separately cannot be provided. In this work, we explore methods to detect which high-level patterns (i.e. programming plans) have been used in a submission, so learners can receive feedback on planning skills even when their code is not completely correct. Our preliminary results show that LLMs with few-shot prompting can detect the use of programming plans in 95% of correct and 86% of partially correct submissions. Incorporating LLMs into grading of open-ended programming exercises can enable more fine-grained feedback to students, even in cases where their code does not compile due to other errors.
KW - autograding
KW - large language models
KW - programming plans
UR - http://www.scopus.com/inward/record.url?scp=86000243309&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=86000243309&partnerID=8YFLogxK
U2 - 10.1145/3641555.3705166
DO - 10.1145/3641555.3705166
M3 - Conference contribution
AN - SCOPUS:86000243309
T3 - SIGCSE TS 2025 - Proceedings of the 56th ACM Technical Symposium on Computer Science Education
SP - 1435
EP - 1436
BT - SIGCSE TS 2025 - Proceedings of the 56th ACM Technical Symposium on Computer Science Education
PB - Association for Computing Machinery
T2 - 56th Annual SIGCSE Technical Symposium on Computer Science Education, SIGCSE TS 2025
Y2 - 26 February 2025 through 1 March 2025
ER -