This study shows that faster (up to three orders of magnitude) SHAP implementations are as accurate as the slower model-agnostic one. As more hidden layers were added to the neural network, the descriptive accuracy increased by about 10%.
It has been found that the effect of deep learning can be best viewed when assessing the trustworthiness of explanation models. In doing so, it evaluates the impact of deep learning (multi-layer perceptron neural networks) on the performance of AES. Building on previous studies where models were trained to predict both the holistic and rubric scores of essays, using the Automated Student Assessment Prize’s essay datasets, this study focuses on predicting the quality of the writing style of Grade-7 essays and exposes the decision processes that lead to these predictions. Although several algorithms from Explainable Artificial Intelligence have recently been published, no research has yet investigated the role that these explanation models can play in: (a) discovering the decision-making process that drives AES, (b) fine-tuning predictive models to improve generalizability and interpretability, and (c) providing personalized, formative, and fine-grained feedback to students during the writing process. Consequently, the AES black box has remained impenetrable. However, a vast swath of research tackles AES only holistically few have even developed AES models at the rubric level, the very first layer of explanation underlying the prediction of holistic scores. School of Computing and Information Systems, Faculty of Science and Technology, Athabasca University, Edmonton, AB, CanadaĪutomated essay scoring (AES) is a compelling topic in Learning Analytics for the primary reason that recent advances in AI find it as a good testbed to explore artificial supplementation of human creativity.