Automated Essay Scoring for Nonnative English Speakers
Article Status
Published
Authors/contributors
- Burstein, Jill (Author)
- Chodorow, Martin (Author)
Title
Automated Essay Scoring for Nonnative English Speakers
Abstract
The e-rater system™ is an operational automated essay scoring system, developed at Educational Testing Service (ETS). The average agreement between human readers, and between independent human readers and e-rater is approximately 92%. There is much interest in the larger writing community in examining the system's performance on nonnative speaker essays. This paper focuses on results of a study that show e-rater's performance on Test of Written English (TWE) essay responses written by nonnative English speakers whose native language is Chinese, Arabic, or Spanish. In addition, one small sample of the data is from US-born English speakers, and another is from non-US-born candidates who report that their native language is English. As expected, significant differences were found among the scores of the English groups and the nonnative speakers. While there were also differences between e-rater and the human readers for the various language groups, the average agreement rate was as high as operational agreement. At least four of the five features that are included in e-rater's current operational models (including discourse, topical, and syntactic features) also appear in the TWE models. This suggests that the features generalize well over a wide range of linguistic variation, as e-rater was not confounded by non-standard English syntactic structures or stylistic discourse structures which one might expect to be a problem for a system designed to evaluate native speaker writing.
Date
1999
Proceedings Title
Proceedings of a Symposium on Computer Mediated Language Assessment and Evaluation in Natural Language Processing - ASSESSEVALNLP '99
Conference Name
a Symposium
Place
College Park, Maryland
Publisher
Association for Computational Linguistics
Pages
68
Series
ASSESSEVALNLP '99
Extra
event-place: College Park, Maryland
Citation Key: burstein1999a
<标题>: 针对非英语母语者的自动作文评分
<AI Smry>: Results of a study that show e-rater's performance on Test of Written English (TWE) essay responses written by nonnative English speakers whose native language is Chinese, Arabic, or Spanish suggest that the features generalize well over a wide range of linguistic variation.
Citation
Burstein, J., & Chodorow, M. (1999). Automated Essay Scoring for Nonnative English Speakers. Proceedings of a Symposium on Computer Mediated Language Assessment and Evaluation in Natural Language Processing - ASSESSEVALNLP ’99, 68. http://portal.acm.org/citation.cfm?doid=1598834.1598847
Link to this record