Short-answer scoring with ensembles of pretrained language models
Article Status
Published
Author/contributor
- Ormerod, Christopher (Author)
Title
Short-answer scoring with ensembles of pretrained language models
Abstract
We investigate the effectiveness of ensembles of pretrained transformer-based language models on short answer questions using the Kaggle Automated Short Answer Scoring dataset. We fine-tune a collection of popular small, base, and large pretrained transformer-based language models, and train one feature-base model on the dataset with the aim of testing ensembles of these models. We used an early stopping mechanism and hyperparameter optimization in training. We observe that generally that the larger models perform slightly better, however, they still fall short of state-of-the-art results one their own. Once we consider ensembles of models, there are ensembles of a number of large networks that do produce state-of-the-art results, however, these ensembles are too large to realistically be put in a production environment.
Repository
arXiv
Archive ID
arXiv:2202.11558
Date
2022-02-23
Accessed
10/05/2024, 02:15
Library Catalogue
Extra
arXiv:2202.11558 [cs]
Citation Key: ormerod2022
Citation
Ormerod, C. (2022). Short-answer scoring with ensembles of pretrained language models (arXiv:2202.11558). arXiv. http://arxiv.org/abs/2202.11558
Link to this record