{"title":"Les Printemps de la Médecine Interne : l’intelligence artificielle face aux experts internistes","authors":"A. Albaladejo , A. Lorleac’h , J.-S. Allain","doi":"10.1016/j.revmed.2024.01.012","DOIUrl":null,"url":null,"abstract":"<div><h3>Introduction</h3><p>The “Printemps de la Médecine Interne” are training days for Francophone internists. The clinical cases presented during these days are complex. This study aims to evaluate the diagnostic capabilities of non-specialized artificial intelligence (language models) ChatGPT-4 and Bard by confronting them with the puzzles of the “Printemps de la Médecine Interne”.</p></div><div><h3>Method</h3><p>Clinical cases from the “Printemps de la Médecine Interne” 2021 and 2022 were submitted to two language models: ChatGPT-4 and Bard. In case of a wrong answer, a second attempt was offered. We then compared the responses of human internist experts to those of artificial intelligence.</p></div><div><h3>Results</h3><p>Of the 12 clinical cases submitted, human internist experts diagnosed nine, ChatGPT-4 diagnosed three, and Bard diagnosed one. One of the cases solved by ChatGPT-4 was not solved by the internist expert. The artificial intelligence had a response time of a few seconds.</p></div><div><h3>Conclusions</h3><p>Currently, the diagnostic skills of ChatGPT-4 and Bard are inferior to those of human experts in solving complex clinical cases but are very promising. Recently made available to the general public, they already have impressive capabilities, questioning the role of the diagnostic physician. It would be advisable to adapt the rules or subjects of future “Printemps de la Médecine Interne” so that they are not solved by a public language model.</p></div>","PeriodicalId":0,"journal":{"name":"","volume":"45 7","pages":"Pages 409-414"},"PeriodicalIF":0.0,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0248866324000328/pdfft?md5=a553b4a34e0c183c8b6d87a043f38a57&pid=1-s2.0-S0248866324000328-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0248866324000328","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Introduction
The “Printemps de la Médecine Interne” are training days for Francophone internists. The clinical cases presented during these days are complex. This study aims to evaluate the diagnostic capabilities of non-specialized artificial intelligence (language models) ChatGPT-4 and Bard by confronting them with the puzzles of the “Printemps de la Médecine Interne”.
Method
Clinical cases from the “Printemps de la Médecine Interne” 2021 and 2022 were submitted to two language models: ChatGPT-4 and Bard. In case of a wrong answer, a second attempt was offered. We then compared the responses of human internist experts to those of artificial intelligence.
Results
Of the 12 clinical cases submitted, human internist experts diagnosed nine, ChatGPT-4 diagnosed three, and Bard diagnosed one. One of the cases solved by ChatGPT-4 was not solved by the internist expert. The artificial intelligence had a response time of a few seconds.
Conclusions
Currently, the diagnostic skills of ChatGPT-4 and Bard are inferior to those of human experts in solving complex clinical cases but are very promising. Recently made available to the general public, they already have impressive capabilities, questioning the role of the diagnostic physician. It would be advisable to adapt the rules or subjects of future “Printemps de la Médecine Interne” so that they are not solved by a public language model.