{"title":"Comparative analysis of artificial intelligence chatbots in orthodontic emergency scenarios: ChatGPT-3.5, ChatGPT-4.0, Copilot, and Gemini.","authors":"Buket Erdem, Mustafa Özcan, Çağla Şar","doi":"10.2319/021825-146.1","DOIUrl":null,"url":null,"abstract":"<p><strong>Objectives: </strong>To evaluate and compare the accuracy of four AI chatbots, ChatGPT-3.5, ChatGPT-4.0, Copilot, and Gemini, in response to orthodontic emergency scenarios.</p><p><strong>Materials and methods: </strong>Forty frequently asked questions related to orthodontic emergencies were posed to the chatbots. These questions were categorized as fixed orthodontic treatment, clear aligner treatment, eating and oral hygiene, pain and discomfort, general concerns, retention, and sports and travel. The responses were evaluated by three orthodontic experts using a five-point Likert scale, and statistical analysis was conducted to assess variations in accuracy across chatbots.</p><p><strong>Results: </strong>Statistical analysis revealed significant differences among the chatbots. Gemini and ChatGPT-4.0 demonstrated the highest accuracy in response to orthodontic emergencies, followed by Copilot, whereas ChatGPT-3.5 had the lowest accuracy scores. Additionally, the \"Fixed Orthodontic Treatment\" category showed a statistically significant difference (P = .043), with Gemini outperforming the other chatbots in this category. However, no statistically significant differences were found in other categories.</p><p><strong>Conclusions: </strong>AI chatbots show potential in providing immediate assistance for orthodontic emergencies, but their accuracy varies across different models and question categories.</p>","PeriodicalId":94224,"journal":{"name":"The Angle orthodontist","volume":" ","pages":""},"PeriodicalIF":3.2000,"publicationDate":"2025-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"The Angle orthodontist","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.2319/021825-146.1","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Objectives: To evaluate and compare the accuracy of four AI chatbots, ChatGPT-3.5, ChatGPT-4.0, Copilot, and Gemini, in response to orthodontic emergency scenarios.
Materials and methods: Forty frequently asked questions related to orthodontic emergencies were posed to the chatbots. These questions were categorized as fixed orthodontic treatment, clear aligner treatment, eating and oral hygiene, pain and discomfort, general concerns, retention, and sports and travel. The responses were evaluated by three orthodontic experts using a five-point Likert scale, and statistical analysis was conducted to assess variations in accuracy across chatbots.
Results: Statistical analysis revealed significant differences among the chatbots. Gemini and ChatGPT-4.0 demonstrated the highest accuracy in response to orthodontic emergencies, followed by Copilot, whereas ChatGPT-3.5 had the lowest accuracy scores. Additionally, the "Fixed Orthodontic Treatment" category showed a statistically significant difference (P = .043), with Gemini outperforming the other chatbots in this category. However, no statistically significant differences were found in other categories.
Conclusions: AI chatbots show potential in providing immediate assistance for orthodontic emergencies, but their accuracy varies across different models and question categories.