Are large language models a useful resource to address common patient concerns on hallux valgus? A readability analysis.

Autor: Hlavinka WJ; Texas A&M School of Medicine, Baylor University Medical Center, Department of Medical Education, 3500 Gaston Avenue, 6-Roberts, Dallas, TX 75246, USA. Electronic address: will_hlavinka15@tamu.edu., Sontam TR; Texas A&M School of Medicine, Baylor University Medical Center, Department of Medical Education, 3500 Gaston Avenue, 6-Roberts, Dallas, TX 75246, USA. Electronic address: tarunsontam@tamu.edu., Gupta A; Texas A&M School of Medicine, Baylor University Medical Center, Department of Medical Education, 3500 Gaston Avenue, 6-Roberts, Dallas, TX 75246, USA. Electronic address: ajgupta@tamu.edu., Croen BJ; Department of Orthopedic Surgery, University of Pennsylvania Health System, 51 N 39th St, Philadelphia, PA 19104, USA. Electronic address: brett.croen@pennmedicine.upenn.edu., Abdullah MS; Department of Orthopedic Surgery, University of Pennsylvania Health System, 51 N 39th St, Philadelphia, PA 19104, USA. Electronic address: mohammed.abdullah@pennmedicine.upenn.edu., Humbyrd CJ; Department of Orthopedic Surgery, University of Pennsylvania Health System, 51 N 39th St, Philadelphia, PA 19104, USA. Electronic address: casey.humbyrd@pennmedicine.upenn.edu.
Jazyk: angličtina
Zdroj: Foot and ankle surgery : official journal of the European Society of Foot and Ankle Surgeons [Foot Ankle Surg] 2024 Aug 06. Date of Electronic Publication: 2024 Aug 06.
DOI: 10.1016/j.fas.2024.08.002
Abstrakt: Background: This study evaluates the accuracy and readability of Google, ChatGPT-3.5, and 4.0 (two versions of an artificial intelligence model) responses to common questions regarding bunion surgery.
Methods: A Google search of "bunionectomy" was performed, and the first ten questions under "People Also Ask" were recorded. ChatGPT-3.5 and 4.0 were asked these ten questions individually, and their answers were analyzed using the Flesch-Kincaid Reading Ease and Gunning-Fog Level algorithms.
Results: When compared to Google, ChatGPT-3.5 and 4.0 had a larger word count with 315 ± 39 words (p < .0001) and 294 ± 39 words (p < .0001), respectively. A significant difference was found between ChatGPT-3.5 and 4.0 compared to Google using Flesch-Kincaid Reading Ease (p < .0001).
Conclusions: Our findings demonstrate that ChatGPT provided significantly lengthier responses than Google and there was a significant difference in reading ease. Both platforms exceeded the seventh to eighth-grade reading level of the U.S.
Level of Evidence: N/A.
Competing Interests: Declaration of Competing Interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
(Copyright © 2024 European Foot and Ankle Society. Published by Elsevier Ltd. All rights reserved.)
Databáze: MEDLINE