The Scientific Knowledge of Bard and ChatGPT in Endocrinology, Diabetes, and Diabetes Technology: Multiple-Choice Questions Examination-Based Performance.

Autor: Meo SA; Department of Physiology, College of Medicine, King Saud University, Riyadh, Saudi Arabia., Al-Khlaiwi T; Department of Physiology, College of Medicine, King Saud University, Riyadh, Saudi Arabia., AbuKhalaf AA; College of Medicine, King Saud University, Riyadh, Saudi Arabia., Meo AS; The School of Medicine, Medical Sciences and Nutrition, University of Aberdeen, Aberdeen, UK., Klonoff DC; Diabetes Research Institute, Mills-Peninsula Medical Center, San Mateo, CA, USA.
Jazyk: angličtina
Zdroj: Journal of diabetes science and technology [J Diabetes Sci Technol] 2023 Oct 05, pp. 19322968231203987. Date of Electronic Publication: 2023 Oct 05.
DOI: 10.1177/19322968231203987
Abstrakt: Background: The present study aimed to investigate the knowledge level of Bard and ChatGPT in the areas of endocrinology, diabetes, and diabetes technology through a multiple-choice question (MCQ) examination format.
Methods: Initially, a 100-MCQ bank was established based on MCQs in endocrinology, diabetes, and diabetes technology. The MCQs were created from physiology, medical textbooks, and academic examination pools in the areas of endocrinology, diabetes, and diabetes technology and academic examination pools. The study team members analyzed the MCQ contents to ensure that they were related to the endocrinology, diabetes, and diabetes technology. The number of MCQs from endocrinology was 50, and that from diabetes and science technology was also 50. The knowledge level of Google's Bard and ChatGPT was assessed with an MCQ-based examination.
Results: In the endocrinology examination section, ChatGPT obtained 29 marks (correct responses) of 50 (58%), and Bard obtained a similar score of 29 of 50 (58%). However, in the diabetes technology examination section, ChatGPT obtained 23 marks of 50 (46%), and Bard obtained 20 marks of 50 (40%). Overall, in the entire three-part examination, ChatGPT obtained 52 marks of 100 (52%), and Bard obtained 49 marks of 100 (49%). ChatGPT obtained slightly more marks than Bard. However, both ChatGPT and Bard did not achieve satisfactory scores in endocrinology or diabetes/technology of at least 60%.
Conclusions: The overall MCQ-based performance of ChatGPT was slightly better than that of Google's Bard. However, both ChatGPT and Bard did not achieve appropriate scores in endocrinology and diabetes/diabetes technology. The study indicates that Bard and ChatGPT have the potential to facilitate medical students and faculty in academic medical education settings, but both artificial intelligence tools need more updated information in the fields of endocrinology, diabetes, and diabetes technology.
Competing Interests: Declaration of Conflicting InterestsThe author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Databáze: MEDLINE