TY - GEN
T1 - Feature Selection Metrics
T2 - 13th International Conference on Educational Data Mining, EDM 2020
AU - Sanyal, Debopam
AU - Bosch, Nigel
AU - Paquette, Luc
N1 - Publisher Copyright:
© 2020 Proceedings of the 13th International Conference on Educational Data Mining, EDM 2020. All rights reserved.
PY - 2020
Y1 - 2020
N2 - Supervised machine learning has become one of the most important methods for developing educational and intelligent tutoring software; it is the backbone of many educational data mining methods for estimating knowledge, emotion, and other aspects of learning. Hence, in order to ensure optimal utilization of computing resources and effective analysis of models, it is essential that researchers know which evaluation metrics are best suited to educational data. In this article, we focus on the problem of wrapper feature selection, where predictors are added to models based on how much they improve model accuracy in terms of a given metric. We compared commonly-used machine learning algorithms including naive Bayes, support vector machines, logistic regression, and random forests on 11 diverse learning-related datasets. We optimized feature selection based on nine different metrics, then evaluated each to address research questions about how effective each metric was in terms of the others (e.g., does optimizing for precision also result in good F1?) as well as calibration (i.e., are predictions produced by models accurate probabilities of correctness?). We provide empirical evidence that the Matthews correlation coefficient (MCC) produced the overall best results across the other metrics, but that root mean squared error (RMSE) selected the best-calibrated models. Finally, we also discuss issues related to the number of features selected when optimizing for each metric, as well as the types of datasets for which certain metrics were more effective.
AB - Supervised machine learning has become one of the most important methods for developing educational and intelligent tutoring software; it is the backbone of many educational data mining methods for estimating knowledge, emotion, and other aspects of learning. Hence, in order to ensure optimal utilization of computing resources and effective analysis of models, it is essential that researchers know which evaluation metrics are best suited to educational data. In this article, we focus on the problem of wrapper feature selection, where predictors are added to models based on how much they improve model accuracy in terms of a given metric. We compared commonly-used machine learning algorithms including naive Bayes, support vector machines, logistic regression, and random forests on 11 diverse learning-related datasets. We optimized feature selection based on nine different metrics, then evaluated each to address research questions about how effective each metric was in terms of the others (e.g., does optimizing for precision also result in good F1?) as well as calibration (i.e., are predictions produced by models accurate probabilities of correctness?). We provide empirical evidence that the Matthews correlation coefficient (MCC) produced the overall best results across the other metrics, but that root mean squared error (RMSE) selected the best-calibrated models. Finally, we also discuss issues related to the number of features selected when optimizing for each metric, as well as the types of datasets for which certain metrics were more effective.
KW - Feature selection
KW - Machine learning
KW - Metrics
KW - Student models
UR - http://www.scopus.com/inward/record.url?scp=85102879961&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85102879961&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85102879961
T3 - Proceedings of the 13th International Conference on Educational Data Mining, EDM 2020
SP - 212
EP - 223
BT - Proceedings of the 13th International Conference on Educational Data Mining, EDM 2020
A2 - Rafferty, Anna N.
A2 - Whitehill, Jacob
A2 - Romero, Cristobal
A2 - Cavalli-Sforza, Violetta
PB - International Educational Data Mining Society
Y2 - 10 July 2020 through 13 July 2020
ER -