Daniel Bahir, Audrey Rostov, Yumna Busool Abu Eta, Shirin Hamed Azzam, David Lockington, Joshua C Teichman, Artemis Matsou, Clara C Chan, Elad Shvartz, Michael Mimouni
{"title":"人工智能与眼科专家:对睑炎患者查询的反应的比较分析。","authors":"Daniel Bahir, Audrey Rostov, Yumna Busool Abu Eta, Shirin Hamed Azzam, David Lockington, Joshua C Teichman, Artemis Matsou, Clara C Chan, Elad Shvartz, Michael Mimouni","doi":"10.1177/11206721251350809","DOIUrl":null,"url":null,"abstract":"<p><p>ObjectiveTo assess the accuracy and clinical education value of responses from AI models (GPT-3.5, GPT-4o, Gemini, Gemini Advanced) compared to expert ophthalmologists' answers to common patient questions about blepharitis, and evaluate their potential for patient education and clinical use.MethodsThirteen frequently asked questions about blepharitis were selected. Responses were generated by AI models and compared to expert answers. A panel of ophthalmologists rated each response for correctness and clinical education value using a 7-point Likert scale. The Friedman test with post hoc comparisons was used to identify performance differences.ResultsExpert responses had the highest correctness (6.3) and clinical education value (6.4) scores, especially in complex, context-driven questions. Significant differences were found between expert and AI responses (<i>P</i> < 0.05). Among AI models, GPT-3.5 performed best in simple definitions (correctness: 6.4) but dropped to 5.5 in nuanced cases. GPT-4o followed (5.4), while Gemini and Gemini Advanced scored lower (5.0 and 4.9), especially in diagnostic and treatment contexts.ConclusionsAI models can support patient education by effectively answering basic factual questions about blepharitis. However, their limitations in complex clinical scenarios highlight the continued need for expert input. While promising as educational tools, AI should complement-not replace-clinician guidance in patient care.</p>","PeriodicalId":12000,"journal":{"name":"European Journal of Ophthalmology","volume":" ","pages":"11206721251350809"},"PeriodicalIF":1.4000,"publicationDate":"2025-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Artificial intelligence versus ophthalmology experts: Comparative analysis of responses to blepharitis patient queries.\",\"authors\":\"Daniel Bahir, Audrey Rostov, Yumna Busool Abu Eta, Shirin Hamed Azzam, David Lockington, Joshua C Teichman, Artemis Matsou, Clara C Chan, Elad Shvartz, Michael Mimouni\",\"doi\":\"10.1177/11206721251350809\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>ObjectiveTo assess the accuracy and clinical education value of responses from AI models (GPT-3.5, GPT-4o, Gemini, Gemini Advanced) compared to expert ophthalmologists' answers to common patient questions about blepharitis, and evaluate their potential for patient education and clinical use.MethodsThirteen frequently asked questions about blepharitis were selected. Responses were generated by AI models and compared to expert answers. A panel of ophthalmologists rated each response for correctness and clinical education value using a 7-point Likert scale. The Friedman test with post hoc comparisons was used to identify performance differences.ResultsExpert responses had the highest correctness (6.3) and clinical education value (6.4) scores, especially in complex, context-driven questions. Significant differences were found between expert and AI responses (<i>P</i> < 0.05). Among AI models, GPT-3.5 performed best in simple definitions (correctness: 6.4) but dropped to 5.5 in nuanced cases. GPT-4o followed (5.4), while Gemini and Gemini Advanced scored lower (5.0 and 4.9), especially in diagnostic and treatment contexts.ConclusionsAI models can support patient education by effectively answering basic factual questions about blepharitis. However, their limitations in complex clinical scenarios highlight the continued need for expert input. While promising as educational tools, AI should complement-not replace-clinician guidance in patient care.</p>\",\"PeriodicalId\":12000,\"journal\":{\"name\":\"European Journal of Ophthalmology\",\"volume\":\" \",\"pages\":\"11206721251350809\"},\"PeriodicalIF\":1.4000,\"publicationDate\":\"2025-06-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"European Journal of Ophthalmology\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1177/11206721251350809\",\"RegionNum\":4,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"OPHTHALMOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"European Journal of Ophthalmology","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1177/11206721251350809","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"OPHTHALMOLOGY","Score":null,"Total":0}
Artificial intelligence versus ophthalmology experts: Comparative analysis of responses to blepharitis patient queries.
ObjectiveTo assess the accuracy and clinical education value of responses from AI models (GPT-3.5, GPT-4o, Gemini, Gemini Advanced) compared to expert ophthalmologists' answers to common patient questions about blepharitis, and evaluate their potential for patient education and clinical use.MethodsThirteen frequently asked questions about blepharitis were selected. Responses were generated by AI models and compared to expert answers. A panel of ophthalmologists rated each response for correctness and clinical education value using a 7-point Likert scale. The Friedman test with post hoc comparisons was used to identify performance differences.ResultsExpert responses had the highest correctness (6.3) and clinical education value (6.4) scores, especially in complex, context-driven questions. Significant differences were found between expert and AI responses (P < 0.05). Among AI models, GPT-3.5 performed best in simple definitions (correctness: 6.4) but dropped to 5.5 in nuanced cases. GPT-4o followed (5.4), while Gemini and Gemini Advanced scored lower (5.0 and 4.9), especially in diagnostic and treatment contexts.ConclusionsAI models can support patient education by effectively answering basic factual questions about blepharitis. However, their limitations in complex clinical scenarios highlight the continued need for expert input. While promising as educational tools, AI should complement-not replace-clinician guidance in patient care.
期刊介绍:
The European Journal of Ophthalmology was founded in 1991 and is issued in print bi-monthly. It publishes only peer-reviewed original research reporting clinical observations and laboratory investigations with clinical relevance focusing on new diagnostic and surgical techniques, instrument and therapy updates, results of clinical trials and research findings.