Assessing ChatGPT's theoretical knowledge and prescriptive accuracy in bacterial infections: a comparative study with infectious diseases residents and specialists.

IF 5.4 2区 医学 Q1 INFECTIOUS DISEASES
Infection Pub Date : 2025-06-01 Epub Date: 2024-07-12 DOI:10.1007/s15010-024-02350-6
Andrea De Vito, Nicholas Geremia, Andrea Marino, Davide Fiore Bavaro, Giorgia Caruana, Marianna Meschiari, Agnese Colpani, Maria Mazzitelli, Vincenzo Scaglione, Emmanuele Venanzi Rullo, Vito Fiore, Marco Fois, Edoardo Campanella, Eugenia Pistarà, Matteo Faltoni, Giuseppe Nunnari, Annamaria Cattelan, Cristina Mussini, Michele Bartoletti, Luigi Angelo Vaira, Giordano Madeddu
{"title":"Assessing ChatGPT's theoretical knowledge and prescriptive accuracy in bacterial infections: a comparative study with infectious diseases residents and specialists.","authors":"Andrea De Vito, Nicholas Geremia, Andrea Marino, Davide Fiore Bavaro, Giorgia Caruana, Marianna Meschiari, Agnese Colpani, Maria Mazzitelli, Vincenzo Scaglione, Emmanuele Venanzi Rullo, Vito Fiore, Marco Fois, Edoardo Campanella, Eugenia Pistarà, Matteo Faltoni, Giuseppe Nunnari, Annamaria Cattelan, Cristina Mussini, Michele Bartoletti, Luigi Angelo Vaira, Giordano Madeddu","doi":"10.1007/s15010-024-02350-6","DOIUrl":null,"url":null,"abstract":"<p><strong>Objectives: </strong>Advancements in Artificial Intelligence(AI) have made platforms like ChatGPT increasingly relevant in medicine. This study assesses ChatGPT's utility in addressing bacterial infection-related questions and antibiogram-based clinical cases.</p><p><strong>Methods: </strong>This study involved a collaborative effort involving infectious disease (ID) specialists and residents. A group of experts formulated six true/false, six open-ended questions, and six clinical cases with antibiograms for four types of infections (endocarditis, pneumonia, intra-abdominal infections, and bloodstream infection) for a total of 96 questions. The questions were submitted to four senior residents and four specialists in ID and inputted into ChatGPT-4 and a trained version of ChatGPT-4. A total of 720 responses were obtained and reviewed by a blinded panel of experts in antibiotic treatments. They evaluated the responses for accuracy and completeness, the ability to identify correct resistance mechanisms from antibiograms, and the appropriateness of antibiotics prescriptions.</p><p><strong>Results: </strong>No significant difference was noted among the four groups for true/false questions, with approximately 70% correct answers. The trained ChatGPT-4 and ChatGPT-4 offered more accurate and complete answers to the open-ended questions than both the residents and specialists. Regarding the clinical case, we observed a lower accuracy from ChatGPT-4 to recognize the correct resistance mechanism. ChatGPT-4 tended not to prescribe newer antibiotics like cefiderocol or imipenem/cilastatin/relebactam, favoring less recommended options like colistin. Both trained- ChatGPT-4 and ChatGPT-4 recommended longer than necessary treatment periods (p-value = 0.022).</p><p><strong>Conclusions: </strong>This study highlights ChatGPT's capabilities and limitations in medical decision-making, specifically regarding bacterial infections and antibiogram analysis. While ChatGPT demonstrated proficiency in answering theoretical questions, it did not consistently align with expert decisions in clinical case management. Despite these limitations, the potential of ChatGPT as a supportive tool in ID education and preliminary analysis is evident. However, it should not replace expert consultation, especially in complex clinical decision-making.</p>","PeriodicalId":13600,"journal":{"name":"Infection","volume":" ","pages":"873-881"},"PeriodicalIF":5.4000,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12137519/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Infection","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1007/s15010-024-02350-6","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/7/12 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"INFECTIOUS DISEASES","Score":null,"Total":0}
引用次数: 0

Abstract

Objectives: Advancements in Artificial Intelligence(AI) have made platforms like ChatGPT increasingly relevant in medicine. This study assesses ChatGPT's utility in addressing bacterial infection-related questions and antibiogram-based clinical cases.

Methods: This study involved a collaborative effort involving infectious disease (ID) specialists and residents. A group of experts formulated six true/false, six open-ended questions, and six clinical cases with antibiograms for four types of infections (endocarditis, pneumonia, intra-abdominal infections, and bloodstream infection) for a total of 96 questions. The questions were submitted to four senior residents and four specialists in ID and inputted into ChatGPT-4 and a trained version of ChatGPT-4. A total of 720 responses were obtained and reviewed by a blinded panel of experts in antibiotic treatments. They evaluated the responses for accuracy and completeness, the ability to identify correct resistance mechanisms from antibiograms, and the appropriateness of antibiotics prescriptions.

Results: No significant difference was noted among the four groups for true/false questions, with approximately 70% correct answers. The trained ChatGPT-4 and ChatGPT-4 offered more accurate and complete answers to the open-ended questions than both the residents and specialists. Regarding the clinical case, we observed a lower accuracy from ChatGPT-4 to recognize the correct resistance mechanism. ChatGPT-4 tended not to prescribe newer antibiotics like cefiderocol or imipenem/cilastatin/relebactam, favoring less recommended options like colistin. Both trained- ChatGPT-4 and ChatGPT-4 recommended longer than necessary treatment periods (p-value = 0.022).

Conclusions: This study highlights ChatGPT's capabilities and limitations in medical decision-making, specifically regarding bacterial infections and antibiogram analysis. While ChatGPT demonstrated proficiency in answering theoretical questions, it did not consistently align with expert decisions in clinical case management. Despite these limitations, the potential of ChatGPT as a supportive tool in ID education and preliminary analysis is evident. However, it should not replace expert consultation, especially in complex clinical decision-making.

Abstract Image

评估 ChatGPT 在细菌感染方面的理论知识和处方准确性:与传染病住院医师和专科医生的比较研究。
目的:人工智能(AI)的进步使得像 ChatGPT 这样的平台在医学领域的应用越来越广泛。本研究评估了 ChatGPT 在解决细菌感染相关问题和基于抗生素图谱的临床病例方面的实用性:本研究由传染病(ID)专家和住院医师共同参与。一组专家针对四种类型的感染(心内膜炎、肺炎、腹腔内感染和血流感染)制定了六道真/假问题、六道开放式问题和六个带有抗生素图谱的临床病例,共计 96 道问题。这些问题提交给了四位资深住院医师和四位内科专家,并输入到 ChatGPT-4 和 ChatGPT-4 的培训版中。抗生素治疗专家组成的盲人小组共收到 720 条回复并进行了审核。他们对回答的准确性和完整性、从抗生素图中识别正确耐药机制的能力以及抗生素处方的适当性进行了评估:结果:在真/假问题上,四组之间没有明显差异,正确率约为 70%。与住院医师和专科医生相比,经过培训的 ChatGPT-4 和 ChatGPT-4 对开放式问题的回答更准确、更完整。关于临床病例,我们观察到 ChatGPT-4 识别正确阻力机制的准确率较低。ChatGPT-4 倾向于不开具头孢克肟或亚胺培南/西司他丁/雷帕坦等较新抗生素的处方,而倾向于开具可乐定等推荐度较低的处方。经过培训的 ChatGPT-4 和 ChatGPT-4 建议的治疗时间都长于必要的治疗时间(p 值 = 0.022):本研究强调了 ChatGPT 在医疗决策方面的能力和局限性,特别是在细菌感染和抗生素图谱分析方面。虽然 ChatGPT 在回答理论问题方面表现出了一定的能力,但在临床病例管理中,它与专家的决策并不一致。尽管存在这些局限性,但 ChatGPT 作为 ID 教育和初步分析的辅助工具的潜力是显而易见的。但是,它不应取代专家咨询,尤其是在复杂的临床决策中。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Infection
Infection 医学-传染病学
CiteScore
12.50
自引率
1.30%
发文量
224
审稿时长
6-12 weeks
期刊介绍: Infection is a journal dedicated to serving as a global forum for the presentation and discussion of clinically relevant information on infectious diseases. Its primary goal is to engage readers and contributors from various regions around the world in the exchange of knowledge about the etiology, pathogenesis, diagnosis, and treatment of infectious diseases, both in outpatient and inpatient settings. The journal covers a wide range of topics, including: Etiology: The study of the causes of infectious diseases. Pathogenesis: The process by which an infectious agent causes disease. Diagnosis: The methods and techniques used to identify infectious diseases. Treatment: The medical interventions and strategies employed to treat infectious diseases. Public Health: Issues of local, regional, or international significance related to infectious diseases, including prevention, control, and management strategies. Hospital Epidemiology: The study of the spread of infectious diseases within healthcare settings and the measures to prevent nosocomial infections. In addition to these, Infection also includes a specialized "Images" section, which focuses on high-quality visual content, such as images, photographs, and microscopic slides, accompanied by brief abstracts. This section is designed to highlight the clinical and diagnostic value of visual aids in the field of infectious diseases, as many conditions present with characteristic clinical signs that can be diagnosed through inspection, and imaging and microscopy are crucial for accurate diagnosis. The journal's comprehensive approach ensures that it remains a valuable resource for healthcare professionals and researchers in the field of infectious diseases.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信