{"title":"革命性的炎症性肠病保健沟通:胃肠病学家和ChatGPT反应的正面比较。","authors":"Zixuan He, Yilong Liu, Zhaoming Wang, Xiaoyu Zhou, Xuanming Fan, Miao He, Chengcheng Wang, Hongyu Fu, Zhijie Wang, Yu Bai","doi":"10.1093/gastro/goaf087","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Artificial intelligence-driven large language models demonstrate immense potential in the medical field. It remains unclear whether ChatGPT has the ability to provide appropriate recommendations for patients with inflammatory bowel disease (IBD) that are comparable to those of gastroenterologists. This study quantitatively assessed the performance of ChatGPT's generated IBD-related recommendations from the distinct perspectives of gastroenterologists and patients.</p><p><strong>Methods: </strong>Healthcare questions regarding IBD were solicited from IBD patients and specialized physicians. Those questions were then presented to GPT-4 Omni and three independent senior gastroenterologists for responses. These responses were subsequently evaluated by a blinded panel of five board-certified gastroenterologists using a five-point Likert scale, assessing accuracy, completeness, and readability. Furthermore, 10 IBD patients as blinded assessors performed assessments of both ChatGPT's and gastroenterologists' responses.</p><p><strong>Results: </strong>Thirty high-frequency questions were selected, encompassing basic knowledge, treatment, and management domains. ChatGPT demonstrated high reproducibility in responding to these questions. Regarding accuracy and readability, ChatGPT's performance was comparable to that of gastroenterologists. For completeness of responses, ChatGPT outperformed gastroenterologists (4.42 ± 0.67 vs 4.19 ± 0.65; <i>P </i>= 0.012). Overall, IBD patients were satisfied with both ChatGPT's and gastroenterologists' responses but, for treatment-related questions, patients rated gastroenterologists higher than ChatGPT (4.54 ± 0.32 vs 4.21 ± 0.38; <i>P </i>= 0.040).</p><p><strong>Conclusions: </strong>ChatGPT has the potential to provide stable, accurate, comprehensive, and comprehensible healthcare-related information for IBD patients. Further validation of the reliability and practicality of large language models in real-world clinical settings is crucial.</p>","PeriodicalId":54275,"journal":{"name":"Gastroenterology Report","volume":"13 ","pages":"goaf087"},"PeriodicalIF":4.2000,"publicationDate":"2025-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12479395/pdf/","citationCount":"0","resultStr":"{\"title\":\"Revolutionizing inflammatory bowel disease healthcare communication: a head-to-head comparison of gastroenterologist and ChatGPT responses.\",\"authors\":\"Zixuan He, Yilong Liu, Zhaoming Wang, Xiaoyu Zhou, Xuanming Fan, Miao He, Chengcheng Wang, Hongyu Fu, Zhijie Wang, Yu Bai\",\"doi\":\"10.1093/gastro/goaf087\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Background: </strong>Artificial intelligence-driven large language models demonstrate immense potential in the medical field. It remains unclear whether ChatGPT has the ability to provide appropriate recommendations for patients with inflammatory bowel disease (IBD) that are comparable to those of gastroenterologists. This study quantitatively assessed the performance of ChatGPT's generated IBD-related recommendations from the distinct perspectives of gastroenterologists and patients.</p><p><strong>Methods: </strong>Healthcare questions regarding IBD were solicited from IBD patients and specialized physicians. Those questions were then presented to GPT-4 Omni and three independent senior gastroenterologists for responses. These responses were subsequently evaluated by a blinded panel of five board-certified gastroenterologists using a five-point Likert scale, assessing accuracy, completeness, and readability. Furthermore, 10 IBD patients as blinded assessors performed assessments of both ChatGPT's and gastroenterologists' responses.</p><p><strong>Results: </strong>Thirty high-frequency questions were selected, encompassing basic knowledge, treatment, and management domains. ChatGPT demonstrated high reproducibility in responding to these questions. Regarding accuracy and readability, ChatGPT's performance was comparable to that of gastroenterologists. For completeness of responses, ChatGPT outperformed gastroenterologists (4.42 ± 0.67 vs 4.19 ± 0.65; <i>P </i>= 0.012). Overall, IBD patients were satisfied with both ChatGPT's and gastroenterologists' responses but, for treatment-related questions, patients rated gastroenterologists higher than ChatGPT (4.54 ± 0.32 vs 4.21 ± 0.38; <i>P </i>= 0.040).</p><p><strong>Conclusions: </strong>ChatGPT has the potential to provide stable, accurate, comprehensive, and comprehensible healthcare-related information for IBD patients. Further validation of the reliability and practicality of large language models in real-world clinical settings is crucial.</p>\",\"PeriodicalId\":54275,\"journal\":{\"name\":\"Gastroenterology Report\",\"volume\":\"13 \",\"pages\":\"goaf087\"},\"PeriodicalIF\":4.2000,\"publicationDate\":\"2025-09-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12479395/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Gastroenterology Report\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1093/gastro/goaf087\",\"RegionNum\":3,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/1/1 0:00:00\",\"PubModel\":\"eCollection\",\"JCR\":\"Q2\",\"JCRName\":\"GASTROENTEROLOGY & HEPATOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Gastroenterology Report","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1093/gastro/goaf087","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/1/1 0:00:00","PubModel":"eCollection","JCR":"Q2","JCRName":"GASTROENTEROLOGY & HEPATOLOGY","Score":null,"Total":0}
引用次数: 0
摘要
背景:人工智能驱动的大型语言模型在医学领域显示出巨大的潜力。目前尚不清楚ChatGPT是否有能力为炎症性肠病(IBD)患者提供与胃肠病学家相当的适当建议。本研究从胃肠病学家和患者的不同角度定量评估了ChatGPT生成的ibd相关建议的性能。方法:向IBD患者和专科医生询问有关IBD的保健问题。然后将这些问题提交给GPT-4 Omni和三位独立的高级胃肠病学家进行回答。随后,由五名经委员会认证的胃肠病学家组成的盲法小组使用五点李克特量表对这些反应进行评估,评估准确性、完整性和可读性。此外,10名IBD患者作为盲法评估者对ChatGPT和胃肠病学家的反应进行了评估。结果:选取了30个高频问题,涵盖基础知识、治疗和管理领域。ChatGPT在回答这些问题时表现出了很高的可重复性。在准确性和可读性方面,ChatGPT的表现与胃肠病学家相当。在回答的完整性方面,ChatGPT优于胃肠病学家(4.42±0.67 vs 4.19±0.65;P = 0.012)。总体而言,IBD患者对ChatGPT和胃肠病学家的回答都很满意,但在治疗相关的问题上,患者对胃肠病学家的评价高于ChatGPT(4.54±0.32 vs 4.21±0.38;P = 0.040)。结论:ChatGPT有可能为IBD患者提供稳定、准确、全面、易懂的医疗相关信息。在现实世界的临床环境中进一步验证大型语言模型的可靠性和实用性至关重要。
Revolutionizing inflammatory bowel disease healthcare communication: a head-to-head comparison of gastroenterologist and ChatGPT responses.
Background: Artificial intelligence-driven large language models demonstrate immense potential in the medical field. It remains unclear whether ChatGPT has the ability to provide appropriate recommendations for patients with inflammatory bowel disease (IBD) that are comparable to those of gastroenterologists. This study quantitatively assessed the performance of ChatGPT's generated IBD-related recommendations from the distinct perspectives of gastroenterologists and patients.
Methods: Healthcare questions regarding IBD were solicited from IBD patients and specialized physicians. Those questions were then presented to GPT-4 Omni and three independent senior gastroenterologists for responses. These responses were subsequently evaluated by a blinded panel of five board-certified gastroenterologists using a five-point Likert scale, assessing accuracy, completeness, and readability. Furthermore, 10 IBD patients as blinded assessors performed assessments of both ChatGPT's and gastroenterologists' responses.
Results: Thirty high-frequency questions were selected, encompassing basic knowledge, treatment, and management domains. ChatGPT demonstrated high reproducibility in responding to these questions. Regarding accuracy and readability, ChatGPT's performance was comparable to that of gastroenterologists. For completeness of responses, ChatGPT outperformed gastroenterologists (4.42 ± 0.67 vs 4.19 ± 0.65; P = 0.012). Overall, IBD patients were satisfied with both ChatGPT's and gastroenterologists' responses but, for treatment-related questions, patients rated gastroenterologists higher than ChatGPT (4.54 ± 0.32 vs 4.21 ± 0.38; P = 0.040).
Conclusions: ChatGPT has the potential to provide stable, accurate, comprehensive, and comprehensible healthcare-related information for IBD patients. Further validation of the reliability and practicality of large language models in real-world clinical settings is crucial.
期刊介绍:
Gastroenterology Report is an international fully open access (OA) online only journal, covering all areas related to gastrointestinal sciences, including studies of the alimentary tract, liver, biliary, pancreas, enteral nutrition and related fields. The journal aims to publish high quality research articles on both basic and clinical gastroenterology, authoritative reviews that bring together new advances in the field, as well as commentaries and highlight pieces that provide expert analysis of topical issues.