Systematic review of ChatGPT accuracy and performance in Iran’s medical licensing exams: A brief report

ChatGPT has demonstrated significant potential in various aspects of medicine, including its performance on licensing examinations. In this study, we systematically investigated ChatGPT’s performance in Iranian medical exams and assessed the quality of the included studies using a previously publish...

Full description

Saved in:
Bibliographic Details
Main Authors: Alireza Keshtkar, Farnaz Atighi, Hamid Reihani
Format: Article
Language:English
Published: Wolters Kluwer Medknow Publications 2024-11-01
Series:Journal of Education and Health Promotion
Subjects:
Online Access:https://journals.lww.com/10.4103/jehp.jehp_1210_24
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:ChatGPT has demonstrated significant potential in various aspects of medicine, including its performance on licensing examinations. In this study, we systematically investigated ChatGPT’s performance in Iranian medical exams and assessed the quality of the included studies using a previously published assessment checklist. The study found that ChatGPT achieved an accuracy range of 32–72% on basic science exams, 34–68.5% on pre-internship exams, and 32–84% on residency exams. Notably, its performance was generally higher when the input was provided in English compared to Persian. One study reported a 40% accuracy rate on an endodontic board exam. To establish ChatGPT as a supplementary tool in medical education and clinical practice, we suggest that dedicated guidelines and checklists are needed to ensure high-quality and consistent research in this emerging field.
ISSN:2277-9531
2319-6440