Abstract
This study applies the state of the art in explainable AI techniques to shed light on the automated essay scoring (AES) process. By means of linear regression and Shapley values, SHAP (Shapley Additive Explanations) approximates a complex AES predictive model implemented as a deep neural network and an ensemble regression. This study delves into the essentials of the automated assessment of ‘organization’, a key rubric in writing. Specifically, it explores whether the organization and connections between ideas and/or events are clear and logically sequenced. Built on findings from previous work, this paper, in addition to improving the generalizability and interpretability of the AES model, highlights the means to identify important ‘writing features’ (both global and local) and hint at the best ranges of feature values. By associating ‘organization’ with ‘writing features’, it provides a mechanism to hypothesize causal relationships among variables and shape machine-learned formative feedback in human-friendly terms for the consumption of teachers and students. Finally, it offers an in-depth discussion on linguistic aspects implied by the findings.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Kumar, V., Boulanger, D.: Automated essay scoring and the deep learning black box: how are rubric scores determined? Int. J. Artif. Intell. Educ. (submitted)
Murdoch, W.J., Singh, C., Kumbier, K., Abbasi-Asl, R., Yu, B.: Definitions, methods, and applications in interpretable machine learning. Proc. Natl. Acad. Sci. 116, 22071–22080 (2019)
Lundberg, S.M., Lee, S.-I.: A unified approach to interpreting model predictions. In: Guyon, I., et al. (eds.) Advances in Neural Information Processing Systems, pp. 4765–4774. (2017)
Kumar, V., Fraser, S.N., Boulanger, D.: Discovering the predictive power of five baseline writing competences. J. Writ. Anal. 1, 176–226 (2017)
Molnar, C.: Interpretable machine learning. Lulu.com (2019). https://christophm.github.io/interpretable-ml-book/
Ribeiro, M.T., Singh, S., Guestrin, C.: “Why should I trust you?”: explaining the predictions of any classifier. CoRR. abs/1602.0 (2016)
Taghipour, K.: Robust trait-specific essay scoring using neural networks and density estimators. Doctoral dissertation (2017)
Zupanc, K., Bosnić, Z.: Automated essay evaluation with semantic analysis. Knowl.-Based Syst. 120, 118–132 (2017)
McCarthy, P.M., Jarvis, S.: MTLD, vocd-D, and HD-D: a validation study of sophisticated approaches to lexical diversity assessment. Behav. Res. Methods 42, 381–392 (2010)
Torruella, J., Capsada, R.: Lexical statistics and typological structures: a measure of lexical richness. Procedia-Soc. Behav. Sci. 95, 447–454 (2013)
Pearl, J., Mackenzie, D.: The Book of Why: The New Science of Cause and Effect. Basic Books, New York (2018)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Boulanger, D., Kumar, V. (2020). SHAPed Automated Essay Scoring: Explaining Writing Features’ Contributions to English Writing Organization. In: Kumar, V., Troussas, C. (eds) Intelligent Tutoring Systems. ITS 2020. Lecture Notes in Computer Science(), vol 12149. Springer, Cham. https://doi.org/10.1007/978-3-030-49663-0_10
Download citation
DOI: https://doi.org/10.1007/978-3-030-49663-0_10
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-49662-3
Online ISBN: 978-3-030-49663-0
eBook Packages: Computer ScienceComputer Science (R0)