TY - GEN
T1 - Intrinsically Interpretable Artificial Neural Networks for Learner Modeling
AU - Pinto, Juan D.
AU - Paquette, Luc
AU - Bosch, Nigel
N1 - Publisher Copyright:
© 2024 Copyright is held by the author(s).
PY - 2024
Y1 - 2024
N2 - Modern AI algorithms are so complex, that it is often impossible even for expert AI engineers to fully explain how they make decisions. Researchers in education are increasingly using such“black-box”algorithms for a wide variety of tasks. This lack of transparency has rightfully raised concerns over issues of fairness, accountability, and trust. Post-hoc explainability techniques exist that aim to address this issue. However, studies in both educational and non-educational contexts have highlighted fundamental problems with these approaches. In this proposed project, we take an alternative approach that aims to make complex AI learner models more intrinsically interpretable, while illustrating how such interpretability can be evaluated. We aim to (1) develop an interpretable neural network, comparing accuracy and issues relevant to interpretability approaches as a whole, (2) evaluate this model’s level of interpretability using a humangrounded evaluation approach, and (3) validate the model’s inner representations and explore some hypothetical advantages of interpretable models, including their use for knowledge discovery.
AB - Modern AI algorithms are so complex, that it is often impossible even for expert AI engineers to fully explain how they make decisions. Researchers in education are increasingly using such“black-box”algorithms for a wide variety of tasks. This lack of transparency has rightfully raised concerns over issues of fairness, accountability, and trust. Post-hoc explainability techniques exist that aim to address this issue. However, studies in both educational and non-educational contexts have highlighted fundamental problems with these approaches. In this proposed project, we take an alternative approach that aims to make complex AI learner models more intrinsically interpretable, while illustrating how such interpretability can be evaluated. We aim to (1) develop an interpretable neural network, comparing accuracy and issues relevant to interpretability approaches as a whole, (2) evaluate this model’s level of interpretability using a humangrounded evaluation approach, and (3) validate the model’s inner representations and explore some hypothetical advantages of interpretable models, including their use for knowledge discovery.
KW - evaluating interpretability
KW - Explainable AI
KW - interpretable neural networks
KW - model transparency
UR - https://www.scopus.com/pages/publications/105023283900
UR - https://www.scopus.com/pages/publications/105023283900#tab=citedBy
U2 - 10.5281/zenodo.12730021
DO - 10.5281/zenodo.12730021
M3 - Conference contribution
AN - SCOPUS:105023283900
SN - 9781733673655
T3 - Proceedings of the International Conference on Educational Data Mining
SP - 982
EP - 985
BT - Proceedings of the 17th International Conference on Educational Data Mining, EDM 2024
A2 - Demmans Epp, Carrie
A2 - Paaßen, Benjamin
A2 - Joyner, David
PB - International Educational Data Mining Society
T2 - 17th International Conference on Educational Data Mining, EDM 2024
Y2 - 14 July 2024 through 17 July 2024
ER -