Artificial intelligence in the prescription of acute medical treatments in primary healthcare - comparison of the performance of family physicians and ChatGPT.
Bárbara Lemos Pereira Simão, Carolina Moura Pereira, Mariana Jácome, Catarina Oliveira, Laura Martins Ferreira, José Miguel Paiva, Carlos Braga, Carlos Seiça Cardoso
{"title":"Artificial intelligence in the prescription of acute medical treatments in primary healthcare - comparison of the performance of family physicians and ChatGPT.","authors":"Bárbara Lemos Pereira Simão, Carolina Moura Pereira, Mariana Jácome, Catarina Oliveira, Laura Martins Ferreira, José Miguel Paiva, Carlos Braga, Carlos Seiça Cardoso","doi":"10.1186/s12875-025-02963-2","DOIUrl":null,"url":null,"abstract":"<p><strong>Introduction: </strong>Artificial intelligence (AI) is increasingly being recognized as a transformative force in healthcare, showing significant promise in supporting healthcare professionals. AI has many applications in healthcare, including providing real-time decision support, diagnosing diseases, and advancing personalized medicine. However, clinical trials and further research are needed to evaluate the practical effectiveness of AI in primary healthcare.</p><p><strong>Objective of the study: </strong>This study aims to assess the accuracy of ChatGPT, an AI-powered chatbot, in therapeutic decision-making during acute disease consultations in primary care and compare its performance to that of general family physicians. The goal was to determine how well ChatGPT could replicate the decisions made by physicians based on standard clinical guidelines.</p><p><strong>Materials and methods: </strong>A cross-sectional study was conducted at three primary healthcare units in the Central Region of Portugal. The analysis involved three phases: (1) collecting data from healthcare professionals, (2) gathering therapeutic proposals from ChatGPT v3.5 based on physician-defined diagnoses, and (3) comparing the treatments proposed by both ChatGPT v3.5 and the physicians, using the Dynamed platform as the gold standard for correct prescriptions.</p><p><strong>Results: </strong>Out of a total of 860 consultations, 138 were excluded due to non-compliance with the inclusion criteria. The analysis showed that the diagnostic accuracy of ChatGPT v3.5 and physicians co-occurred in 26.2% of cases. In 29.1% of cases, there was no agreement between the AI and the physicians' diagnoses. The therapeutic decisions made by ChatGPT v3.5 were correct in 55.6% of the cases, while physicians made correct decisions in 54.3% of the cases. The therapeutic decisions of ChatGPT v3.5 were incorrect in 5.2% of the cases, compared to 11% for physicians. Furthermore, the therapeutic proposals of ChatGPT v3.5 were 'approximate' to the correct treatment in 24% of the cases, while physicians had a 17.1% approximation rate.</p><p><strong>Conclusion: </strong>This study suggests that AI - specifically ChatGPT v3.5 - can match or even outperform physicians in terms of therapeutic decision accuracy, with a similar or slightly better success rate than human doctors. This highlights the potential for AI to act as an effective auxiliary tool rather than a replacement for healthcare professionals. AI is most effective when used in collaboration with healthcare professionals, augmenting their capabilities and improving overall healthcare delivery. Ultimately, AI can serve as a powerful aid to healthcare professionals, helping improve patient care and healthcare outcomes, particularly in primary care.</p>","PeriodicalId":72428,"journal":{"name":"BMC primary care","volume":"26 1","pages":"284"},"PeriodicalIF":2.6000,"publicationDate":"2025-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12455777/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"BMC primary care","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1186/s12875-025-02963-2","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"MEDICINE, GENERAL & INTERNAL","Score":null,"Total":0}
引用次数: 0
Abstract
Introduction: Artificial intelligence (AI) is increasingly being recognized as a transformative force in healthcare, showing significant promise in supporting healthcare professionals. AI has many applications in healthcare, including providing real-time decision support, diagnosing diseases, and advancing personalized medicine. However, clinical trials and further research are needed to evaluate the practical effectiveness of AI in primary healthcare.
Objective of the study: This study aims to assess the accuracy of ChatGPT, an AI-powered chatbot, in therapeutic decision-making during acute disease consultations in primary care and compare its performance to that of general family physicians. The goal was to determine how well ChatGPT could replicate the decisions made by physicians based on standard clinical guidelines.
Materials and methods: A cross-sectional study was conducted at three primary healthcare units in the Central Region of Portugal. The analysis involved three phases: (1) collecting data from healthcare professionals, (2) gathering therapeutic proposals from ChatGPT v3.5 based on physician-defined diagnoses, and (3) comparing the treatments proposed by both ChatGPT v3.5 and the physicians, using the Dynamed platform as the gold standard for correct prescriptions.
Results: Out of a total of 860 consultations, 138 were excluded due to non-compliance with the inclusion criteria. The analysis showed that the diagnostic accuracy of ChatGPT v3.5 and physicians co-occurred in 26.2% of cases. In 29.1% of cases, there was no agreement between the AI and the physicians' diagnoses. The therapeutic decisions made by ChatGPT v3.5 were correct in 55.6% of the cases, while physicians made correct decisions in 54.3% of the cases. The therapeutic decisions of ChatGPT v3.5 were incorrect in 5.2% of the cases, compared to 11% for physicians. Furthermore, the therapeutic proposals of ChatGPT v3.5 were 'approximate' to the correct treatment in 24% of the cases, while physicians had a 17.1% approximation rate.
Conclusion: This study suggests that AI - specifically ChatGPT v3.5 - can match or even outperform physicians in terms of therapeutic decision accuracy, with a similar or slightly better success rate than human doctors. This highlights the potential for AI to act as an effective auxiliary tool rather than a replacement for healthcare professionals. AI is most effective when used in collaboration with healthcare professionals, augmenting their capabilities and improving overall healthcare delivery. Ultimately, AI can serve as a powerful aid to healthcare professionals, helping improve patient care and healthcare outcomes, particularly in primary care.