Conversational Guide for Cataract Surgery Complications: A Comparative Study of Surgeons versus Large Language Model-Based Chatbot Generated Instructions for Patient Interaction.

IF 1.7 4区 医学 Q3 OPHTHALMOLOGY
Sathishkumar Sundaramoorthy, Vineet Ratra, Vijay Shankar, Ramesh Dorairajan, Quresh Maskati, T Nirmal Fredrick, Aashna Ratra, Dhanashree Ratra
{"title":"Conversational Guide for Cataract Surgery Complications: A Comparative Study of Surgeons versus Large Language Model-Based Chatbot Generated Instructions for Patient Interaction.","authors":"Sathishkumar Sundaramoorthy, Vineet Ratra, Vijay Shankar, Ramesh Dorairajan, Quresh Maskati, T Nirmal Fredrick, Aashna Ratra, Dhanashree Ratra","doi":"10.1080/09286586.2025.2484772","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>It is difficult to explain the complications of surgery to patients. Care has to be taken to convey the facts clearly and objectively while expressing concern for their wellbeing. This study compared responses from surgeons with responses from a large language model (LLM)-based chatbot.</p><p><strong>Methods: </strong>We presented 10 common scenarios of cataract surgery complications to seven senior surgeons and a chatbot. The responses were graded by two independent graders for comprehension, readability, and complexity of language using previously validated indices. The responses were analyzed for accuracy and completeness. Honesty and empathy were graded for both groups. Scores were averaged and tabulated.</p><p><strong>Results: </strong>The readability scores for the surgeons (10.64) were significantly less complex than the chatbot (12.54) (<i>p</i> < 0.001). The responses from the surgeons were shorter, whereas the chatbot tended to give more detailed answers. The average accuracy and completeness score of chatbot-generated conversations was 2.36 (0.55), which was similar to the surgeons' score of 2.58 (0.36) (<i>p</i> = 0.164). The responses from the chatbot were more generalized, lacking specific alternative measures. While empathy scores were higher for surgeons (1.81 vs. 1.20, <i>p</i> = 0.041), honesty scores showed no significant difference.</p><p><strong>Conclusions: </strong>The LLM-based chatbot gave a detailed description of the complication but was less specific about the alternative measures. The surgeons had a more in-depth understanding of the situation. The chatbot showed complete honesty but scored less for empathy. With more training using complex real-world scenarios and specialized ophthalmologic data, the chatbots could be used to assist the surgeons in counselling patients for postoperative complications.</p>","PeriodicalId":19607,"journal":{"name":"Ophthalmic epidemiology","volume":" ","pages":"1-8"},"PeriodicalIF":1.7000,"publicationDate":"2025-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Ophthalmic epidemiology","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1080/09286586.2025.2484772","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"OPHTHALMOLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

Purpose: It is difficult to explain the complications of surgery to patients. Care has to be taken to convey the facts clearly and objectively while expressing concern for their wellbeing. This study compared responses from surgeons with responses from a large language model (LLM)-based chatbot.

Methods: We presented 10 common scenarios of cataract surgery complications to seven senior surgeons and a chatbot. The responses were graded by two independent graders for comprehension, readability, and complexity of language using previously validated indices. The responses were analyzed for accuracy and completeness. Honesty and empathy were graded for both groups. Scores were averaged and tabulated.

Results: The readability scores for the surgeons (10.64) were significantly less complex than the chatbot (12.54) (p < 0.001). The responses from the surgeons were shorter, whereas the chatbot tended to give more detailed answers. The average accuracy and completeness score of chatbot-generated conversations was 2.36 (0.55), which was similar to the surgeons' score of 2.58 (0.36) (p = 0.164). The responses from the chatbot were more generalized, lacking specific alternative measures. While empathy scores were higher for surgeons (1.81 vs. 1.20, p = 0.041), honesty scores showed no significant difference.

Conclusions: The LLM-based chatbot gave a detailed description of the complication but was less specific about the alternative measures. The surgeons had a more in-depth understanding of the situation. The chatbot showed complete honesty but scored less for empathy. With more training using complex real-world scenarios and specialized ophthalmologic data, the chatbots could be used to assist the surgeons in counselling patients for postoperative complications.

白内障手术并发症的会话指南:外科医生与基于大语言模型的聊天机器人生成的患者交互指令的比较研究。
目的:向患者解释手术并发症是困难的。在表达对他们福祉的关心的同时,必须注意清楚和客观地传达事实。这项研究比较了外科医生的反应和基于大型语言模型(LLM)的聊天机器人的反应。方法:我们向7位资深外科医生和一个聊天机器人介绍10种常见的白内障手术并发症。回答是由两个独立的评分者评分的理解,可读性和语言的复杂性使用先前验证的指标。分析了回答的准确性和完整性。两组的诚实度和同理心都被打分。将分数取平均值并制成表格。结果:外科医生的可读性评分(10.64)明显低于聊天机器人(12.54)(p p = 0.164)。聊天机器人的回答更笼统,缺乏具体的替代措施。外科医生共情得分较高(1.81比1.20,p = 0.041),诚实得分无显著差异。结论:基于llm的聊天机器人给出了并发症的详细描述,但对替代措施不太具体。外科医生对情况有了更深入的了解。聊天机器人表现出完全的诚实,但在同理心方面得分较低。通过使用复杂的真实场景和专业的眼科数据进行更多的训练,聊天机器人可以帮助外科医生为患者提供术后并发症咨询。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Ophthalmic epidemiology
Ophthalmic epidemiology 医学-眼科学
CiteScore
3.70
自引率
5.60%
发文量
61
审稿时长
6-12 weeks
期刊介绍: Ophthalmic Epidemiology is dedicated to the publication of original research into eye and vision health in the fields of epidemiology, public health and the prevention of blindness. Ophthalmic Epidemiology publishes editorials, original research reports, systematic reviews and meta-analysis articles, brief communications and letters to the editor on all subjects related to ophthalmic epidemiology. A broad range of topics is suitable, such as: evaluating the risk of ocular diseases, general and specific study designs, screening program implementation and evaluation, eye health care access, delivery and outcomes, therapeutic efficacy or effectiveness, disease prognosis and quality of life, cost-benefit analysis, biostatistical theory and risk factor analysis. We are looking to expand our engagement with reports of international interest, including those regarding problems affecting developing countries, although reports from all over the world potentially are suitable. Clinical case reports, small case series (not enough for a cohort analysis) articles and animal research reports are not appropriate for this journal.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信