人工智能聊天机器人ChatGPT-3.5、ChatGPT-4.0、Copilot和Gemini在正畸急救场景中的对比分析

IF 3.2
Buket Erdem, Mustafa Özcan, Çağla Şar
{"title":"人工智能聊天机器人ChatGPT-3.5、ChatGPT-4.0、Copilot和Gemini在正畸急救场景中的对比分析","authors":"Buket Erdem, Mustafa Özcan, Çağla Şar","doi":"10.2319/021825-146.1","DOIUrl":null,"url":null,"abstract":"<p><strong>Objectives: </strong>To evaluate and compare the accuracy of four AI chatbots, ChatGPT-3.5, ChatGPT-4.0, Copilot, and Gemini, in response to orthodontic emergency scenarios.</p><p><strong>Materials and methods: </strong>Forty frequently asked questions related to orthodontic emergencies were posed to the chatbots. These questions were categorized as fixed orthodontic treatment, clear aligner treatment, eating and oral hygiene, pain and discomfort, general concerns, retention, and sports and travel. The responses were evaluated by three orthodontic experts using a five-point Likert scale, and statistical analysis was conducted to assess variations in accuracy across chatbots.</p><p><strong>Results: </strong>Statistical analysis revealed significant differences among the chatbots. Gemini and ChatGPT-4.0 demonstrated the highest accuracy in response to orthodontic emergencies, followed by Copilot, whereas ChatGPT-3.5 had the lowest accuracy scores. Additionally, the \"Fixed Orthodontic Treatment\" category showed a statistically significant difference (P = .043), with Gemini outperforming the other chatbots in this category. However, no statistically significant differences were found in other categories.</p><p><strong>Conclusions: </strong>AI chatbots show potential in providing immediate assistance for orthodontic emergencies, but their accuracy varies across different models and question categories.</p>","PeriodicalId":94224,"journal":{"name":"The Angle orthodontist","volume":" ","pages":""},"PeriodicalIF":3.2000,"publicationDate":"2025-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Comparative analysis of artificial intelligence chatbots in orthodontic emergency scenarios: ChatGPT-3.5, ChatGPT-4.0, Copilot, and Gemini.\",\"authors\":\"Buket Erdem, Mustafa Özcan, Çağla Şar\",\"doi\":\"10.2319/021825-146.1\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Objectives: </strong>To evaluate and compare the accuracy of four AI chatbots, ChatGPT-3.5, ChatGPT-4.0, Copilot, and Gemini, in response to orthodontic emergency scenarios.</p><p><strong>Materials and methods: </strong>Forty frequently asked questions related to orthodontic emergencies were posed to the chatbots. These questions were categorized as fixed orthodontic treatment, clear aligner treatment, eating and oral hygiene, pain and discomfort, general concerns, retention, and sports and travel. The responses were evaluated by three orthodontic experts using a five-point Likert scale, and statistical analysis was conducted to assess variations in accuracy across chatbots.</p><p><strong>Results: </strong>Statistical analysis revealed significant differences among the chatbots. Gemini and ChatGPT-4.0 demonstrated the highest accuracy in response to orthodontic emergencies, followed by Copilot, whereas ChatGPT-3.5 had the lowest accuracy scores. Additionally, the \\\"Fixed Orthodontic Treatment\\\" category showed a statistically significant difference (P = .043), with Gemini outperforming the other chatbots in this category. However, no statistically significant differences were found in other categories.</p><p><strong>Conclusions: </strong>AI chatbots show potential in providing immediate assistance for orthodontic emergencies, but their accuracy varies across different models and question categories.</p>\",\"PeriodicalId\":94224,\"journal\":{\"name\":\"The Angle orthodontist\",\"volume\":\" \",\"pages\":\"\"},\"PeriodicalIF\":3.2000,\"publicationDate\":\"2025-09-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"The Angle orthodontist\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.2319/021825-146.1\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"The Angle orthodontist","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.2319/021825-146.1","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

目的:评估和比较四种AI聊天机器人ChatGPT-3.5、ChatGPT-4.0、Copilot和Gemini在正畸紧急情况下的准确性。材料与方法:向聊天机器人提出40个与正畸紧急情况相关的常见问题。这些问题被分类为固定正畸治疗、透明矫正器治疗、饮食和口腔卫生、疼痛和不适、一般问题、滞留、运动和旅行。三位正畸专家使用五点李克特量表对这些回答进行了评估,并进行了统计分析,以评估聊天机器人在准确性方面的差异。结果:统计分析显示聊天机器人之间存在显著差异。Gemini和ChatGPT-4.0在应对正畸紧急情况时表现出最高的准确性,其次是Copilot,而ChatGPT-3.5的准确性得分最低。此外,“固定正畸治疗”类别显示出统计学上显著的差异(P = 0.043), Gemini在这一类别中的表现优于其他聊天机器人。然而,在其他类别中没有发现统计学上的显著差异。结论:人工智能聊天机器人显示出在正畸紧急情况下提供即时援助的潜力,但其准确性因模型和问题类别而异。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Comparative analysis of artificial intelligence chatbots in orthodontic emergency scenarios: ChatGPT-3.5, ChatGPT-4.0, Copilot, and Gemini.

Objectives: To evaluate and compare the accuracy of four AI chatbots, ChatGPT-3.5, ChatGPT-4.0, Copilot, and Gemini, in response to orthodontic emergency scenarios.

Materials and methods: Forty frequently asked questions related to orthodontic emergencies were posed to the chatbots. These questions were categorized as fixed orthodontic treatment, clear aligner treatment, eating and oral hygiene, pain and discomfort, general concerns, retention, and sports and travel. The responses were evaluated by three orthodontic experts using a five-point Likert scale, and statistical analysis was conducted to assess variations in accuracy across chatbots.

Results: Statistical analysis revealed significant differences among the chatbots. Gemini and ChatGPT-4.0 demonstrated the highest accuracy in response to orthodontic emergencies, followed by Copilot, whereas ChatGPT-3.5 had the lowest accuracy scores. Additionally, the "Fixed Orthodontic Treatment" category showed a statistically significant difference (P = .043), with Gemini outperforming the other chatbots in this category. However, no statistically significant differences were found in other categories.

Conclusions: AI chatbots show potential in providing immediate assistance for orthodontic emergencies, but their accuracy varies across different models and question categories.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信