Department of Medicine, School of Medicine, The University of Auckland, Auckland, New Zealand
© 2025 Korea Health Personnel Licensing Examination Institute
This is an open-access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Authors’ contributions
All the work was done by Albert Andrew.
Conflict of interest
No potential conflict of interest relevant to this article was reported.
Funding
None.
Data availability
Data files are available from Harvard Dataverse: https://doi.org/10.7910/DVN/MYQKNY
Dataset 1. ChatGPT cardiology board exam accuracy data for text-based questions.
Dataset 2. ChatGPT cardiology board exam accuracy data for image-based questions.
Acknowledgments
None.
First author surname | Year of study | Version(s) of ChatGPT assessed | Description of question set | Overall reported accuracy |
---|---|---|---|---|
Skalidis et al. [2] | 2023 | ChatGPT (unspecified) | The dataset includes a total of 362 text-based multiple-choice questions sourced from various materials. This includes 68 questions derived from European Society of Cardiology sample resources, 144 questions from the 2022 edition of StudyPRN test questions, and another 150 questions from Braunwald’s Heart Disease Review and Assessment textbook. | Text-based accuracy: ChatGPT-3.5=58.8% (n=312/362) |
Milutinovic et al. [3] | 2024 | ChatGPT-3.5 and ChatGPT-4.0 | 98 Text-based multiple-choice questions were chosen from a selected Cardiovascular Medicine chapter within the Medical Knowledge Self-Assessment Program (MKSAP-19). | Text-based accuracy: ChatGPT-3.5=64.3% (n=63/98); ChatGPT-4.0=74.5% (n=73/98) |
Gritti et al. [4] | 2024 (a) | ChatGPT-3.5 Plus and ChatGPT-4.0 Plus | 88 Text-based multiple-choice questions from the Paediatric Cardiology Board Review textbook | Text-based accuracy: ChatGPT-3.5 Plus=37.5% (n=33/88); ChatGPT-4.0 Plus=65.9% (n=58/88) |
Builoff et al. [5] | 2024 | ChatGPT-4, ChatGPT-4 Turbo, and ChatGPT-4omni (GPT-4o) | 168 Multiple-choice questions (141 text-only and 27 image-based) from the 2023 American Society of Nuclear Cardiology Board Preparation Exam | Text-based accuracy: ChatGPT-4.0=59.2%; ChatGPT-4 Turbo=62.4%; ChatGPT-4omni (GPT-4o)=66.7% |
Image based accuracy: ChatGPT-4.0=40.7%; ChatGPT-4 Turbo=44.4%; ChatGPT-4omni (GPT-4o)=44.4% | ||||
Huwiler et al. [6] | 2024 | ChatGPT-3.5, ChatGPT-3.5 Plus, and ChatGPT-4.0 Plus | The dataset features 88 multiple-choice questions from the Switzerland Cardiological Board Exam. Of these, 76 were text-based questions and 12 were image-based questions. These questions cover various cardiology topics and subfields and are based on the 10th Edition of Braunwald’s Heart Disease Review and Assessment. | Text-based accuracy: ChatGPT-3.5=46.05% (n=35/76); ChatGPT-3.5 Plus=51.30% (n=39/76); ChatGPT-4.0 Plus=76.3% (n=58/76) |
Image-based accuracy: unable to be determined | ||||
Gritti et al. [7] | 2024 (b) | ChatGPT-4.0 Plus | 100 Multiple-choice questions with and without accompanying images from the Paediatric Cardiology Board Review textbook | Image based accuracy: ChatGPT-4.0 Plus=41.0% (n=41/100) |
Text-based accuracy: ChatGPT-4.0 Plus=37.0% (n=37/100) | ||||
Alexandrou et al. [8] | 2024 | ChatGPT-4.0 | 60 Multiple-choice questions were included from the CATHSAP platform. This resource simulates the American College of Cardiology/Society for Cardiovascular Angiography and Interventions international cardiology certification examination, providing a robust preparation tool for candidates | Text-based accuracy: unable to be determined |
Image based accuracy: ChaGPT-4.0=83.3% (n=5/6) |
First author surname | Year of study | Version(s) of ChatGPT assessed | Description of question set | Overall reported accuracy |
---|---|---|---|---|
Skalidis et al. [2] | 2023 | ChatGPT (unspecified) | The dataset includes a total of 362 text-based multiple-choice questions sourced from various materials. This includes 68 questions derived from European Society of Cardiology sample resources, 144 questions from the 2022 edition of StudyPRN test questions, and another 150 questions from Braunwald’s Heart Disease Review and Assessment textbook. | Text-based accuracy: ChatGPT-3.5=58.8% (n=312/362) |
Milutinovic et al. [3] | 2024 | ChatGPT-3.5 and ChatGPT-4.0 | 98 Text-based multiple-choice questions were chosen from a selected Cardiovascular Medicine chapter within the Medical Knowledge Self-Assessment Program (MKSAP-19). | Text-based accuracy: ChatGPT-3.5=64.3% (n=63/98); ChatGPT-4.0=74.5% (n=73/98) |
Gritti et al. [4] | 2024 (a) | ChatGPT-3.5 Plus and ChatGPT-4.0 Plus | 88 Text-based multiple-choice questions from the Paediatric Cardiology Board Review textbook | Text-based accuracy: ChatGPT-3.5 Plus=37.5% (n=33/88); ChatGPT-4.0 Plus=65.9% (n=58/88) |
Builoff et al. [5] | 2024 | ChatGPT-4, ChatGPT-4 Turbo, and ChatGPT-4omni (GPT-4o) | 168 Multiple-choice questions (141 text-only and 27 image-based) from the 2023 American Society of Nuclear Cardiology Board Preparation Exam | Text-based accuracy: ChatGPT-4.0=59.2%; ChatGPT-4 Turbo=62.4%; ChatGPT-4omni (GPT-4o)=66.7% |
Image based accuracy: ChatGPT-4.0=40.7%; ChatGPT-4 Turbo=44.4%; ChatGPT-4omni (GPT-4o)=44.4% | ||||
Huwiler et al. [6] | 2024 | ChatGPT-3.5, ChatGPT-3.5 Plus, and ChatGPT-4.0 Plus | The dataset features 88 multiple-choice questions from the Switzerland Cardiological Board Exam. Of these, 76 were text-based questions and 12 were image-based questions. These questions cover various cardiology topics and subfields and are based on the 10th Edition of Braunwald’s Heart Disease Review and Assessment. | Text-based accuracy: ChatGPT-3.5=46.05% (n=35/76); ChatGPT-3.5 Plus=51.30% (n=39/76); ChatGPT-4.0 Plus=76.3% (n=58/76) |
Image-based accuracy: unable to be determined | ||||
Gritti et al. [7] | 2024 (b) | ChatGPT-4.0 Plus | 100 Multiple-choice questions with and without accompanying images from the Paediatric Cardiology Board Review textbook | Image based accuracy: ChatGPT-4.0 Plus=41.0% (n=41/100) |
Text-based accuracy: ChatGPT-4.0 Plus=37.0% (n=37/100) | ||||
Alexandrou et al. [8] | 2024 | ChatGPT-4.0 | 60 Multiple-choice questions were included from the CATHSAP platform. This resource simulates the American College of Cardiology/Society for Cardiovascular Angiography and Interventions international cardiology certification examination, providing a robust preparation tool for candidates | Text-based accuracy: unable to be determined |
Image based accuracy: ChaGPT-4.0=83.3% (n=5/6) |