Question Generation by Transformers

Article Status
Published
Authors/contributors
Title
Question Generation by Transformers
Abstract
A machine learning model was developed to automatically generate questions from Wikipedia passages using transformers, an attention-based model eschewing the paradigm of existing recurrent neural networks (RNNs). The model was trained on the inverted Stanford Question Answering Dataset (SQuAD), which is a reading comprehension dataset consisting of 100,000+ questions posed by crowdworkers on a set of Wikipedia articles. After training, the question generation model is able to generate simple questions relevant to unseen passages and answers containing an average of 8 words per question. The word error rate (WER) was used as a metric to compare the similarity between SQuAD questions and the model-generated questions. Although the high average WER suggests that the questions generated differ from the original SQuAD questions, the questions generated are mostly grammatically correct and plausible in their own right.
Repository
arXiv
Archive ID
arXiv:1909.05017
Date
2019-09-14
Citation Key
kriangchaivech2019
Accessed
08/10/2025, 23:11
Library Catalogue
Extra
arXiv:1909.05017 [cs]
Citation
Kriangchaivech, K., & Wangperawong, A. (2019). Question Generation by Transformers (arXiv:1909.05017). arXiv. https://doi.org/10.48550/arXiv.1909.05017
Powered by Zotero and Kerko.