Comparative Performance of ChatGPT and Bard in a Text-Based Radiology Knowledge Assessment

43Citations
Citations of this article
105Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Purpose: Bard by Google, a direct competitor to ChatGPT, was recently released. Understanding the relative performance of these different chatbots can provide important insight into their strengths and weaknesses as well as which roles they are most suited to fill. In this project, we aimed to compare the most recent version of ChatGPT, ChatGPT-4, and Bard by Google, in their ability to accurately respond to radiology board examination practice questions. Methods: Text-based questions were collected from the 2017-2021 American College of Radiology’s Diagnostic Radiology In-Training (DXIT) examinations. ChatGPT-4 and Bard were queried, and their comparative accuracies, response lengths, and response times were documented. Subspecialty-specific performance was analyzed as well. Results: 318 questions were included in our analysis. ChatGPT answered significantly more accurately than Bard (87.11% vs 70.44%, P

Cite

CITATION STYLE

APA

Patil, N. S., Huang, R. S., van der Pol, C. B., & Larocque, N. (2024). Comparative Performance of ChatGPT and Bard in a Text-Based Radiology Knowledge Assessment. Canadian Association of Radiologists Journal, 75(2), 344–350. https://doi.org/10.1177/08465371231193716

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free