Comparative Analysis of Accuracy, Readability, Sentiment, and Actionability: Artificial Intelligence Chatbots (ChatGPT and Google Gemini) versus Traditional Patient Information Leaflets for Local Anesthesia in Eye Surgery.

Autor: Gondode P; Department of Anaesthesiology, Pain medicine and Critical Care, All India Institute of Medical Sciences (AIIMS), New Delhi, India., Duggal S; Department of Anaesthesiology, Pain medicine and Critical Care, All India Institute of Medical Sciences (AIIMS), New Delhi, India., Garg N; Department of Anaesthesiology, Pain medicine and Critical Care, All India Institute of Medical Sciences (AIIMS), New Delhi, India., Lohakare P; Department of Microbiology, Mahatma Gandhi Institute of Medical Sciences (MGIMS), Wardha, India., Jakhar J; Department of Anaesthesiology, Pain medicine and Critical Care, University college of Medical Sciences (UCMS), Delhi, India., Bharti S; Department of Anaesthesiology, Pain medicine and Critical Care, All India Institute of Medical Sciences (AIIMS), New Delhi, India., Dewangan S; Department of Anaesthesiology, Pain medicine and Critical Care, All India Institute of Medical Sciences (AIIMS), New Delhi, India.
Jazyk: angličtina
Zdroj: The British and Irish orthoptic journal [Br Ir Orthopt J] 2024 Aug 19; Vol. 20 (1), pp. 183-192. Date of Electronic Publication: 2024 Aug 19 (Print Publication: 2024).
DOI: 10.22599/bioj.377
Abstrakt: Background and Aim: Eye surgeries often evoke strong negative emotions in patients, including fear and anxiety. Patient education material plays a crucial role in informing and empowering individuals. Traditional sources of medical information may not effectively address individual patient concerns or cater to varying levels of understanding. This study aims to conduct a comparative analysis of the accuracy, completeness, readability, tone, and understandability of patient education material generated by AI chatbots versus traditional Patient Information Leaflets (PILs), focusing on local anesthesia in eye surgery.
Methods: Expert reviewers evaluated responses generated by AI chatbots (ChatGPT and Google Gemini) and a traditional PIL (Royal College of Anaesthetists' PIL) based on accuracy, completeness, readability, sentiment, and understandability. Statistical analyses, including ANOVA and Tukey HSD tests, were conducted to compare the performance of the sources.
Results: Readability analysis showed variations in complexity among the sources, with AI chatbots offering simplified language and PILs maintaining better overall readability and accessibility. Sentiment analysis revealed differences in emotional tone, with Google Gemini exhibiting the most positive sentiment. AI chatbots demonstrated superior understandability and actionability, while PILs excelled in completeness. Overall, ChatGPT showed slightly higher accuracy (scores expressed as mean ± standard deviation) (4.71 ± 0.5 vs 4.61 ± 0.62) and completeness (4.55 ± 0.58 vs 4.47 ± 0.58) compared to Google Gemini, but PILs performed best (4.84 ± 0.37 vs 4.88 ± 0.33) in terms of both accuracy and completeness (p-value for completeness <0.05).
Conclusion: AI chatbots show promise as innovative tools for patient education, complementing traditional PILs. By leveraging the strengths of both AI-driven technologies and human expertise, healthcare providers can enhance patient education and empower individuals to make informed decisions about their health and medical care.
Competing Interests: The authors have no competing interests to declare.
(Copyright: © 2024 The Author(s).)
Databáze: MEDLINE