More efficient processes for creating automated essay scoring frameworks: A demonstration of two algorithms

Article Status
Published
Authors/contributors
Title
More efficient processes for creating automated essay scoring frameworks: A demonstration of two algorithms
Abstract
Automated essay scoring (AES) has emerged as a secondary or as a sole marker for many high-stakes educational assessments, in native and non-native testing, owing to remarkable advances in feature engineering using natural language processing, machine learning, and deep-neural algorithms. The purpose of this study is to compare the effectiveness and the performance of two AES frameworks, each based on machine learning with deep language features, or complex language features, and deep neural algorithms. More specifically, support vector machines (SVMs) in conjunction with Coh-Metrix features were used for a traditional AES model development, and the convolutional neural networks (CNNs) approach was used for more contemporary deep-neural model development. Then, the strengths and weaknesses of the traditional and contemporary models under different circumstances (e.g., types of the rubric, length of the essay, and the essay type) were tested. The results were evaluated using the quadratic weighted kappa (QWK) score and compared with the agreement between the human raters. The results indicated that the CNNs model performs better, meaning that it produced more comparable results to the human raters than the Coh-Metrix + SVMs model. Moreover, the CNNs model also achieved state-of-the-art performance in most of the essay sets with a high average QWK score.
Publication
Language Testing
Volume
38
Issue
2
Pages
247-272
Date
2020-7-4
Journal Abbr
Lang. Test.
Language
en
ISSN
0265-5322
Short Title
More efficient processes for creating automated essay scoring frameworks
Accessed
14/06/2024, 16:22
Library Catalogue
DOI.org (Crossref)
Extra
Citation Key: shin2021 <标题>: 用于创建自动化作文评分框架的更高效流程:两种算法的演示 <AI Smry>: The CNNs model performs better, meaning that it produced more comparable results to the human raters than the Coh-Metrix + SVMs model, and achieved state-of-the-art performance in most of the essay sets with a high average QWK score.
Citation
Shin, J., & Gierl, M. J. (2020). More efficient processes for creating automated essay scoring frameworks: A demonstration of two algorithms. Language Testing, 38(2), 247–272. https://doi.org/10.1177/0265532220937830
Powered by Zotero and Kerko.