Comparing the Quality of Human and ChatGPT Feedback on Students’ Writing

Article Status
Published
Authors/contributors
Title
Comparing the Quality of Human and ChatGPT Feedback on Students’ Writing
Abstract
Offering students formative feedback on drafts of their writing is an effective way to facilitate writing development. This study examined the ability of generative AI (i.e., ChatGPT) to provide formative feedback on students’ compositions. We compared the quality of human and AI feedback by scoring the feedback each provided on secondary student essays (n=200) on five measures of feedback quality: the degree to which feedback (a) was criteria-based, (b) provided clear directions for improvement, (c) was accurate, (d) prioritized essential features, and (e) used a supportive tone. We examined whether ChatGPT and human evaluators provided feedback that differed in quality for native English speakers and English learners and for compositions that differed in overall quality. Results showed that human raters were better at providing high-quality feedback to students in all categories other than criteria-based. Considering the ease of generating feedback through ChatGPT and its overall quality, practical differences between humans and ChatGPT were not substantial. Feedback did not vary by language status for humans or AI, but AI and humans showed differences in feedback based on essay quality. Implications for generative AI as an educational tool are discussed.
Repository
EdArXiv
Date
2023-9-7
Accessed
12/09/2023, 22:07
Language
en-us
Library Catalogue
OSF Preprints
Citation
Steiss, J., Tate, T., Graham, S., Cruz, J., Hebert, M., Wang, J., Moon, Y., Tseng, W., & Uci, M. W. (2023). Comparing the Quality of Human and ChatGPT Feedback on Students’ Writing. EdArXiv. https://doi.org/10.35542/osf.io/ty3em
Powered by Zotero and Kerko.