Abstract
Background: OpenAI released versions ChatGPT-3.5 and GPT-4 between 2022 and 2023. GPT-3.5 has demonstrated proficiency in various examinations, particularly the United States Medical Licensing Examination. However, GPT-4 has more advanced capabilities. Objective: This study aims to examine the efficacy of GPT-3.5 and GPT-4 within the Taiwan National Pharmacist Licensing Examination and to ascertain their utility and potential application in clinical pharmacy and education. Methods: The pharmacist examination in Taiwan consists of 2 stages: basic subjects and clinical subjects. In this study, exam questions were manually fed into the GPT-3.5 and GPT-4 models, and their responses were recorded; graphic-based questions were excluded. This study encompassed three steps: (1) determining the answering accuracy of GPT-3.5 and GPT-4, (2) categorizing question types and observing differences in model performance across these categories, and (3) comparing model performance on calculation and situational questions. Microsoft Excel and R software were used for statistical analyses. Results: GPT-4 achieved an accuracy rate of 72.9%, overshadowing GPT-3.5, which achieved 59.1% (P
Author supplied keywords
- AI
- ChatGPT
- Chatbot
- GPT-4
- Taiwan
- Taiwan national pharmacist licensing examination
- artificial intelligence
- chat generative pre-trained transformer
- educational measurement
- evaluation and comparison study
- generative AI
- learning model
- machine learning
- medical databases
- medical decision-making
- medical education
- pharmacist
- pharmacy
- pharmacy licensure
- statistical analyses
Cite
CITATION STYLE
Wang, Y. M., Shen, H. W., Chen, T. J., Chiang, S. C., & Lin, T. G. (2025). Performance of ChatGPT-3.5 and ChatGPT-4 in the Taiwan National Pharmacist Licensing Examination: Comparative Evaluation Study. JMIR Medical Education, 11. https://doi.org/10.2196/56850
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.