评估人工智能聊天机器人在下颌牵张成骨术术前和术后咨询中的应用。

IF 1 4区 医学 Q3 SURGERY
Taylor Kring, Srihitha Akula, Soumil Prasad, Eric Sokhn, Seth R Thaller
{"title":"评估人工智能聊天机器人在下颌牵张成骨术术前和术后咨询中的应用。","authors":"Taylor Kring, Srihitha Akula, Soumil Prasad, Eric Sokhn, Seth R Thaller","doi":"10.1097/SCS.0000000000011543","DOIUrl":null,"url":null,"abstract":"<p><p>Mandibular distraction osteogenesis (MDO) is a craniofacial procedure frequently performed in pediatric patients with micrognathia airway obstruction. Preoperative and postoperative counseling for families undergoing this procedure is essential involving a multistage surgical course, device management, feeding changes, and airway considerations. This study evaluates the trustworthiness and readability of AI (artificial intelligence) chatbot responses to questions about operative care for MDO. Study was conducted using ChatGPT, Google Gemini, Microsoft Copilot, and Open Evidence. Twenty common preoperative and postoperative care questions relating to MDO were developed. The authors used a modified DISCERN tool to assess quality and the SMOG (Simple Measure of Gobbledygook) test to evaluate response readability. Data underwent statistical analysis using descriptive statistics, 1-way ANOVA, and Tukey HSD. Modified DISCERN analysis revealed clear aims and relevancy scored the highest (mean=4.92 SD=0.31; mean=4.64, SD=0.62). Additional sources provided and citation of sources had the lowest means (mean=2.19 SD=1.52; mean=2.93 SD=1.96). Microsoft Copilot scored the highest in overall quality (mean=38.10 versus ChatGPT=29.90, P<0.001). Open Evidence scored lowest in shared decision-making (mean=1.80 SD=1.10). Effect sizes were large for source-related variables, with eta-squared values >0.75. Significant differences in readability across all AI models were found (mean=17.31 SD=3.59, P<0.001), indicating the average response was at a graduate school reading level. Open Evidence (mean=22.24) produced higher SMOG reading scores than ChatGPT (mean=15.89), Google Gemini (mean=15.66), and Microsoft Copilot (mean=15.44) (P<0.001). These findings highlight a need for reviewing the reliability of AI chatbots in preoperative and postoperative counseling for MDO.</p>","PeriodicalId":15462,"journal":{"name":"Journal of Craniofacial Surgery","volume":" ","pages":""},"PeriodicalIF":1.0000,"publicationDate":"2025-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Evaluating AI Chatbots for Preoperative and Postoperative Counseling for Mandibular Distraction Osteogenesis.\",\"authors\":\"Taylor Kring, Srihitha Akula, Soumil Prasad, Eric Sokhn, Seth R Thaller\",\"doi\":\"10.1097/SCS.0000000000011543\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Mandibular distraction osteogenesis (MDO) is a craniofacial procedure frequently performed in pediatric patients with micrognathia airway obstruction. Preoperative and postoperative counseling for families undergoing this procedure is essential involving a multistage surgical course, device management, feeding changes, and airway considerations. This study evaluates the trustworthiness and readability of AI (artificial intelligence) chatbot responses to questions about operative care for MDO. Study was conducted using ChatGPT, Google Gemini, Microsoft Copilot, and Open Evidence. Twenty common preoperative and postoperative care questions relating to MDO were developed. The authors used a modified DISCERN tool to assess quality and the SMOG (Simple Measure of Gobbledygook) test to evaluate response readability. Data underwent statistical analysis using descriptive statistics, 1-way ANOVA, and Tukey HSD. Modified DISCERN analysis revealed clear aims and relevancy scored the highest (mean=4.92 SD=0.31; mean=4.64, SD=0.62). Additional sources provided and citation of sources had the lowest means (mean=2.19 SD=1.52; mean=2.93 SD=1.96). Microsoft Copilot scored the highest in overall quality (mean=38.10 versus ChatGPT=29.90, P<0.001). Open Evidence scored lowest in shared decision-making (mean=1.80 SD=1.10). Effect sizes were large for source-related variables, with eta-squared values >0.75. Significant differences in readability across all AI models were found (mean=17.31 SD=3.59, P<0.001), indicating the average response was at a graduate school reading level. Open Evidence (mean=22.24) produced higher SMOG reading scores than ChatGPT (mean=15.89), Google Gemini (mean=15.66), and Microsoft Copilot (mean=15.44) (P<0.001). These findings highlight a need for reviewing the reliability of AI chatbots in preoperative and postoperative counseling for MDO.</p>\",\"PeriodicalId\":15462,\"journal\":{\"name\":\"Journal of Craniofacial Surgery\",\"volume\":\" \",\"pages\":\"\"},\"PeriodicalIF\":1.0000,\"publicationDate\":\"2025-06-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Craniofacial Surgery\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1097/SCS.0000000000011543\",\"RegionNum\":4,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"SURGERY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Craniofacial Surgery","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1097/SCS.0000000000011543","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"SURGERY","Score":null,"Total":0}
引用次数: 0

摘要

下颌牵张成骨术(MDO)是一种颅面手术,常用于小颌畸形患儿气道阻塞。对接受该手术的家庭进行术前和术后咨询是必要的,包括多阶段手术过程、设备管理、喂养改变和气道考虑。本研究评估了AI(人工智能)聊天机器人对MDO手术护理问题的回答的可信度和可读性。研究使用ChatGPT、谷歌Gemini、Microsoft Copilot和Open Evidence进行。制定了与MDO相关的20个常见术前和术后护理问题。作者使用改进的DISCERN工具来评估质量,并使用SMOG (Simple Measure of Gobbledygook)测试来评估回答的可读性。数据采用描述性统计、单因素方差分析和Tukey HSD进行统计分析。改进的DISCERN分析显示目标明确,相关性得分最高(平均=4.92,SD=0.31;= 4.64, SD = 0.62)。提供的其他来源和引用的来源的平均值最低(mean=2.19 SD=1.52;意味着= 2.93 SD = 1.96)。微软Copilot在整体质量上得分最高(平均值=38.10,ChatGPT=29.90, P0.75)。所有AI模型的可读性存在显著差异(mean=17.31 SD=3.59, P
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Evaluating AI Chatbots for Preoperative and Postoperative Counseling for Mandibular Distraction Osteogenesis.

Mandibular distraction osteogenesis (MDO) is a craniofacial procedure frequently performed in pediatric patients with micrognathia airway obstruction. Preoperative and postoperative counseling for families undergoing this procedure is essential involving a multistage surgical course, device management, feeding changes, and airway considerations. This study evaluates the trustworthiness and readability of AI (artificial intelligence) chatbot responses to questions about operative care for MDO. Study was conducted using ChatGPT, Google Gemini, Microsoft Copilot, and Open Evidence. Twenty common preoperative and postoperative care questions relating to MDO were developed. The authors used a modified DISCERN tool to assess quality and the SMOG (Simple Measure of Gobbledygook) test to evaluate response readability. Data underwent statistical analysis using descriptive statistics, 1-way ANOVA, and Tukey HSD. Modified DISCERN analysis revealed clear aims and relevancy scored the highest (mean=4.92 SD=0.31; mean=4.64, SD=0.62). Additional sources provided and citation of sources had the lowest means (mean=2.19 SD=1.52; mean=2.93 SD=1.96). Microsoft Copilot scored the highest in overall quality (mean=38.10 versus ChatGPT=29.90, P<0.001). Open Evidence scored lowest in shared decision-making (mean=1.80 SD=1.10). Effect sizes were large for source-related variables, with eta-squared values >0.75. Significant differences in readability across all AI models were found (mean=17.31 SD=3.59, P<0.001), indicating the average response was at a graduate school reading level. Open Evidence (mean=22.24) produced higher SMOG reading scores than ChatGPT (mean=15.89), Google Gemini (mean=15.66), and Microsoft Copilot (mean=15.44) (P<0.001). These findings highlight a need for reviewing the reliability of AI chatbots in preoperative and postoperative counseling for MDO.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
1.70
自引率
11.10%
发文量
968
审稿时长
1.5 months
期刊介绍: ​The Journal of Craniofacial Surgery serves as a forum of communication for all those involved in craniofacial surgery, maxillofacial surgery and pediatric plastic surgery. Coverage ranges from practical aspects of craniofacial surgery to the basic science that underlies surgical practice. The journal publishes original articles, scientific reviews, editorials and invited commentary, abstracts and selected articles from international journals, and occasional international bibliographies in craniofacial surgery.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信