TY - GEN
T1 - An LLM-Based Framework for Simulating, Classifying, and Correcting Students' Programming Knowledge with the SOLO Taxonomy
AU - Zhang, Shan
AU - Meshram, Pragati Shuddhodhan
AU - Ganapathy Prasad, Priyadharshini
AU - Israel, Maya
AU - Bhat, Suma
N1 - This study is supported by the National Science Foundation and the Institute of Education Sciences under Grant DRL-2229612.
PY - 2025/2/18
Y1 - 2025/2/18
N2 - Novice programmers often face challenges in designing computational artifacts and fixing code errors, which can lead to task abandonment and over-reliance on external support. While research has explored effective meta-cognitive strategies to scaffold novice programmers' learning, it is essential to first understand and assess students' conceptual, procedural, and strategic/conditional programming knowledge at scale. To address this issue, we propose a three-model framework that leverages Large Language Models (LLMs) to simulate, classify, and correct student responses to programming questions based on the SOLO Taxonomy. The SOLO Taxonomy provides a structured approach for categorizing student understanding into four levels: Pre-structural, Uni-structural, Multi-structural, and Relational. Our results showed that GPT-4o achieved high accuracy in generating and classifying responses for the Relational category, with moderate accuracy in the Uni-structural and Pre-structural categories, but struggled with the Multi-structural category. The model successfully corrected responses to the Relational level. Although further refinement is needed, these findings suggest that LLMs hold significant potential for supporting computer science education by assessing programming knowledge and guiding students toward deeper cognitive engagement.
AB - Novice programmers often face challenges in designing computational artifacts and fixing code errors, which can lead to task abandonment and over-reliance on external support. While research has explored effective meta-cognitive strategies to scaffold novice programmers' learning, it is essential to first understand and assess students' conceptual, procedural, and strategic/conditional programming knowledge at scale. To address this issue, we propose a three-model framework that leverages Large Language Models (LLMs) to simulate, classify, and correct student responses to programming questions based on the SOLO Taxonomy. The SOLO Taxonomy provides a structured approach for categorizing student understanding into four levels: Pre-structural, Uni-structural, Multi-structural, and Relational. Our results showed that GPT-4o achieved high accuracy in generating and classifying responses for the Relational category, with moderate accuracy in the Uni-structural and Pre-structural categories, but struggled with the Multi-structural category. The model successfully corrected responses to the Relational level. Although further refinement is needed, these findings suggest that LLMs hold significant potential for supporting computer science education by assessing programming knowledge and guiding students toward deeper cognitive engagement.
KW - Computer Science Education
KW - Large Language Model
KW - Solo Taxonomy
UR - http://www.scopus.com/inward/record.url?scp=86000262359&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=86000262359&partnerID=8YFLogxK
U2 - 10.1145/3641555.3705125
DO - 10.1145/3641555.3705125
M3 - Conference contribution
AN - SCOPUS:86000262359
T3 - SIGCSE TS 2025 - Proceedings of the 56th ACM Technical Symposium on Computer Science Education
SP - 1681
EP - 1682
BT - SIGCSE TS 2025 - Proceedings of the 56th ACM Technical Symposium on Computer Science Education
PB - Association for Computing Machinery
T2 - 56th Annual SIGCSE Technical Symposium on Computer Science Education, SIGCSE TS 2025
Y2 - 26 February 2025 through 1 March 2025
ER -