Performance of an Artificial Intelligence-Based Chatbot (ChatGPT) Answering the European Certification in Implant Dentistry Exam.

Marta Revilla-León, Basir A Barmak, Irena Sailer, John C Kois, Wael Att
{"title":"Performance of an Artificial Intelligence-Based Chatbot (ChatGPT) Answering the European Certification in Implant Dentistry Exam.","authors":"Marta Revilla-León, Basir A Barmak, Irena Sailer, John C Kois, Wael Att","doi":"10.11607/ijp.8852","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>To compare the performance of licensed dentists and two software versions (3.5 legacy and 4.0) of an artificial intelligence (AI)-based chatbot (ChatGPT) answering the exam for the 2022 Certification in Implant Dentistry of the European Association for Osseointegration (EAO).</p><p><strong>Materials and methods: </strong>The 50-question, multiple-choice exam of the EAO for the 2022 Certification in Implant Dentistry was obtained. Three groups were created based on the individual or program answering the exam: licensed dentists (D group) and two software versions of an artificial intelligence (AI)-based chatbot (ChatGPT)-3.5 legacy (ChatGPT-3.5 group) and the 4.0 version (ChatGPT-4.0 group). The EAO provided the results of the 2022 examinees (D group). For the ChatGPT groups, the 50 multiple-choice questions were introduced into both ChatGBT versions, and the answers were recorded. Pearson correlation matrix was used to analyze the linear relationship among the subgroups. The inter- and intraoperator reliability was calculated using Cronbach's alpha coefficient. One-way ANOVA and Tukey post-hoc tests were used to examine the data (α = .05).</p><p><strong>Results: </strong>ChatGPT was able to pass the exam for the 2022 Certification in Implant Dentistry of the EAO. Additionally, the software version of ChatGPT impacted the score obtained. The 4.0 version not only pass the exam but also obtained a significantly higher score than the 3.5 version and licensed dentists completing the same exam.</p><p><strong>Conclusions: </strong>The AIbased chatbot tested not only passed the exam but performed better than licensed dentists.</p>","PeriodicalId":94232,"journal":{"name":"The International journal of prosthodontics","volume":"0 0","pages":"221-224"},"PeriodicalIF":0.0000,"publicationDate":"2024-04-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"The International journal of prosthodontics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.11607/ijp.8852","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Purpose: To compare the performance of licensed dentists and two software versions (3.5 legacy and 4.0) of an artificial intelligence (AI)-based chatbot (ChatGPT) answering the exam for the 2022 Certification in Implant Dentistry of the European Association for Osseointegration (EAO).

Materials and methods: The 50-question, multiple-choice exam of the EAO for the 2022 Certification in Implant Dentistry was obtained. Three groups were created based on the individual or program answering the exam: licensed dentists (D group) and two software versions of an artificial intelligence (AI)-based chatbot (ChatGPT)-3.5 legacy (ChatGPT-3.5 group) and the 4.0 version (ChatGPT-4.0 group). The EAO provided the results of the 2022 examinees (D group). For the ChatGPT groups, the 50 multiple-choice questions were introduced into both ChatGBT versions, and the answers were recorded. Pearson correlation matrix was used to analyze the linear relationship among the subgroups. The inter- and intraoperator reliability was calculated using Cronbach's alpha coefficient. One-way ANOVA and Tukey post-hoc tests were used to examine the data (α = .05).

Results: ChatGPT was able to pass the exam for the 2022 Certification in Implant Dentistry of the EAO. Additionally, the software version of ChatGPT impacted the score obtained. The 4.0 version not only pass the exam but also obtained a significantly higher score than the 3.5 version and licensed dentists completing the same exam.

Conclusions: The AIbased chatbot tested not only passed the exam but performed better than licensed dentists.

基于人工智能的聊天机器人 (ChatGPT) 在回答欧洲种植牙认证考试时的表现。
目的:比较执业牙医和两个基于人工智能(AI)的聊天机器人(ChatGPT)软件版本(3.5 legacy 和 4.0)在回答欧洲骨结合协会(EAO)2022 年种植牙认证考试时的表现:获得了欧洲骨整合协会(EAO)2022 年种植牙认证考试的 50 道选择题。根据作答的个人或程序分为三组:持证牙医(D 组)和基于人工智能(AI)的聊天机器人(ChatGPT)的两个软件版本--3.5 版本(ChatGPT-3.5 组)和 4.0 版本(ChatGPT-4.0 组)。EAO 提供了 2022 名受试者(D 组)的结果。对于 ChatGPT 组,在两个 ChatGBT 版本中都引入了 50 道选择题,并记录了答案。使用皮尔逊相关矩阵分析了各分组之间的线性关系。使用 Cronbach's α 系数计算了操作员之间和操作员内部的信度。采用单因素方差分析和 Tukey 事后检验来检验数据(α = .05):结果:ChatGPT 能够通过 EAO 2022 年种植牙认证考试。此外,ChatGPT 的软件版本也影响了得分。4.0 版本不仅能通过考试,而且得分明显高于 3.5 版本和完成相同考试的执业牙医:结论:测试的人工智能聊天机器人不仅能通过考试,而且成绩比执业牙医更好。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信