Use of artificial intelligence-generated multiple-choice questions for the examination of surgical subspecialty residents

Report of feasibility and psychometric analysis

Authors

  • Jin Kyu Kim University of Toronto
  • Michael E. Chua Hospital for Sick Children, Toronto, ON
  • Armando J. Lorenzo The Hospital for Sick Children, Toronto, ON
  • Mandy Rickard Hospital for Sick Children, Toronto, ON
  • Laura Andreacchi University of Toronto
  • Michael Kim University of Toronto
  • Douglas Cheung University of Toronto
  • Yonah Krakowsky University of Toronto
  • Jason Y. Lee University of Toronto

DOI:

https://doi.org/10.5489/cuaj.9020

Keywords:

ChatGPT, Large Language Model, LLM, MCQ, Examination

Abstract

INTRODUCTION: Multiple-choice questions (MCQs) are essential in medical education and widely used by licensing bodies. They are traditionally created with intensive human effort to ensure validity. Recent advances in artificial intelligence (AI), particularly large language models (LLMs), offer the potential to streamline this process. This study aimed to develop and test a GPT-4 model with customized instructions for generating MCQs to assess urology residents.

METHODS: A GPT-4 model was embedded using guidelines from medical licensing bodies and reference materials specific to urology. This model was tasked with generating MCQs designed to mimic the format and content of the 2023 urology examination outlined by the Royal College of Physicians and Surgeons of Canada (RCPSC). Following generation, a selection of MCQs underwent expert review for validity and suitability.

RESULTS: From an initial set of 123 generated MCQs, 60 were chosen for inclusion in an exam administered to 15 urology residents at the University of Toronto. The exam results demonstrated a general increasing performance with level of training cohorts, suggesting the MCQs’ ability to effectively discriminate knowledge levels among residents. The majority (33/60) of the questions had discriminatory value that appeared acceptable (discriminatory index 0.2-0.4) or excellent (discriminatory index >0.4).

CONCLUSIONS: This study highlights AI-driven models like GPT-4 as efficient tools to aid with MCQ generation in medical education assessments. By automating MCQ creation while maintaining quality standards, AI can expedite processes. Future research should focus on refining AI applications in education to optimize assessments and enhance medical training and certification outcomes.

Downloads

Download data is not yet available.

Published

2025-02-24

How to Cite

Kim, J. K., Chua, M. E., Lorenzo, A. . J., Rickard, M. ., Andreacchi, L. ., Kim, M. ., … Lee, J. Y. . (2025). Use of artificial intelligence-generated multiple-choice questions for the examination of surgical subspecialty residents: Report of feasibility and psychometric analysis. Canadian Urological Association Journal, 19(6), 182–7. https://doi.org/10.5489/cuaj.9020

Issue

Section

Original Research