Investigating the comparative superiority of artificial intelligence programs in assessing knowledge levels regarding ocular inflammation, uvea diseases, and treatment modalities.

IF 1 Q4 OPHTHALMOLOGY
Taiwan Journal of Ophthalmology Pub Date : 2024-09-13 eCollection Date: 2024-07-01 DOI:10.4103/tjo.TJO-D-23-00166
Eyupcan Sensoy, Mehmet Citirik
{"title":"Investigating the comparative superiority of artificial intelligence programs in assessing knowledge levels regarding ocular inflammation, uvea diseases, and treatment modalities.","authors":"Eyupcan Sensoy, Mehmet Citirik","doi":"10.4103/tjo.TJO-D-23-00166","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>The purpose of the study was to evaluate the knowledge level of the Chat Generative Pretrained Transformer (ChatGPT), Bard, and Bing artificial intelligence (AI) chatbots regarding ocular inflammation, uveal diseases, and treatment modalities, and to investigate their relative performance compared to one another.</p><p><strong>Materials and methods: </strong>Thirty-six questions related to ocular inflammation, uveal diseases, and treatment modalities were posed to the ChatGPT, Bard, and Bing AI chatbots, and both correct and incorrect responses were recorded. The accuracy rates were compared using the Chi-squared test.</p><p><strong>Results: </strong>The ChatGPT provided correct answers to 52.8% of the questions, while Bard answered 38.9% correctly, and Bing answered 44.4% correctly. All three AI programs provided identical responses to 20 (55.6%) of the questions, with 45% of these responses being correct and 55% incorrect. No significant difference was observed between the correct and incorrect responses from the three AI chatbots (<i>P</i> = 0.654).</p><p><strong>Conclusion: </strong>AI chatbots should be developed to provide widespread access to accurate information about ocular inflammation, uveal diseases, and treatment modalities. Future research could explore ways to enhance the performance of these chatbots.</p>","PeriodicalId":44978,"journal":{"name":"Taiwan Journal of Ophthalmology","volume":"14 3","pages":"409-413"},"PeriodicalIF":1.0000,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11488809/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Taiwan Journal of Ophthalmology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.4103/tjo.TJO-D-23-00166","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/7/1 0:00:00","PubModel":"eCollection","JCR":"Q4","JCRName":"OPHTHALMOLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

Purpose: The purpose of the study was to evaluate the knowledge level of the Chat Generative Pretrained Transformer (ChatGPT), Bard, and Bing artificial intelligence (AI) chatbots regarding ocular inflammation, uveal diseases, and treatment modalities, and to investigate their relative performance compared to one another.

Materials and methods: Thirty-six questions related to ocular inflammation, uveal diseases, and treatment modalities were posed to the ChatGPT, Bard, and Bing AI chatbots, and both correct and incorrect responses were recorded. The accuracy rates were compared using the Chi-squared test.

Results: The ChatGPT provided correct answers to 52.8% of the questions, while Bard answered 38.9% correctly, and Bing answered 44.4% correctly. All three AI programs provided identical responses to 20 (55.6%) of the questions, with 45% of these responses being correct and 55% incorrect. No significant difference was observed between the correct and incorrect responses from the three AI chatbots (P = 0.654).

Conclusion: AI chatbots should be developed to provide widespread access to accurate information about ocular inflammation, uveal diseases, and treatment modalities. Future research could explore ways to enhance the performance of these chatbots.

研究人工智能程序在评估眼部炎症、葡萄膜疾病和治疗方法相关知识水平方面的比较优势。
目的:本研究的目的是评估 Chat Generative Pretrained Transformer(ChatGPT)、Bard 和 Bing 人工智能(AI)聊天机器人对眼部炎症、葡萄膜疾病和治疗方式的了解程度,并研究它们之间的相对性能比较:向 ChatGPT、Bard 和 Bing 人工智能聊天机器人提出了 36 个与眼部炎症、葡萄膜疾病和治疗方式有关的问题,并记录了正确和错误的回答。使用卡方检验比较了正确率:结果:ChatGPT 提供了 52.8% 的正确答案,Bard 回答了 38.9% 的正确答案,Bing 回答了 44.4% 的正确答案。所有三个人工智能程序都对 20 个问题(55.6%)做出了相同的回答,其中 45% 回答正确,55% 回答错误。三个人工智能聊天机器人的正确回答和错误回答之间没有明显差异(P = 0.654):结论:应开发人工智能聊天机器人,以广泛提供有关眼部炎症、葡萄膜疾病和治疗方法的准确信息。未来的研究可以探索提高这些聊天机器人性能的方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
1.80
自引率
9.10%
发文量
68
审稿时长
19 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信