ChatGPT prompts for generating multiple-choice questions in medical education and evidence on their validity: a literature review

Article Status
Published
Authors/contributors
Title
ChatGPT prompts for generating multiple-choice questions in medical education and evidence on their validity: a literature review
Abstract
Abstract ChatGPT’s role in creating multiple-choice questions (MCQs) is growing but the validity of these artificial-intelligence-generated questions is unclear. This literature review was conducted to address the urgent need for understanding the application of ChatGPT in generating MCQs for medical education. Following the database search and screening of 1920 studies, we found 23 relevant studies. We extracted the prompts for MCQ generation and assessed the validity evidence of MCQs. The findings showed that prompts varied, including referencing specific exam styles and adopting specific personas, which align with recommended prompt engineering tactics. The validity evidence covered various domains, showing mixed accuracy rates, with some studies indicating comparable quality to human-written questions, and others highlighting differences in difficulty and discrimination levels, alongside a significant reduction in question creation time. Despite its efficiency, we highlight the necessity of careful review and suggest a need for further research to optimize the use of ChatGPT in question generation. Main messages  Ensure high-quality outputs by utilizing well-designed prompts; medical educators should prioritize the use of detailed, clear ChatGPT prompts when generating MCQs. Avoid using ChatGPT-generated MCQs directly in examinations without thorough review to prevent inaccuracies and ensure relevance. Leverage ChatGPT’s potential to streamline the test development process, enhancing efficiency without compromising quality.
Publication
Postgraduate Medical Journal
Volume
100
Issue
1189
Pages
858-865
Date
2024-10-18
Journal Abbr
Postgrad. Med. J.
Language
en
ISSN
0032-5473, 1469-0756
Short Title
ChatGPT prompts for generating multiple-choice questions in medical education and evidence on their validity
Accessed
08/10/2025, 23:10
Library Catalogue
DOI.org (Crossref)
Extra
Citation Key: kiyak2024
Citation
Kıyak, Y. S., & Emekli, E. (2024). ChatGPT prompts for generating multiple-choice questions in medical education and evidence on their validity: a literature review. Postgraduate Medical Journal, 100(1189), 858–865. https://doi.org/10.1093/postmj/qgae065
Powered by Zotero and Kerko.