Assessing the performance of an artificial intelligence based chatbot in the differential diagnosis of oral mucosal lesions: clinical validation study.
{"title":"Assessing the performance of an artificial intelligence based chatbot in the differential diagnosis of oral mucosal lesions: clinical validation study.","authors":"Nadav Grinberg, Sara Whitefield, Shlomi Kleinman, Clariel Ianculovici, Gilad Wasserman, Oren Peleg","doi":"10.1007/s00784-025-06268-7","DOIUrl":null,"url":null,"abstract":"<p><strong>Objectives: </strong>Artificial intelligence (AI) is becoming more popular in medicine. The current study aims to investigate, primarily, if an AI-based chatbot, such as ChatGPT, could be a valid tool for assisting in establishing a differential diagnosis of oral mucosal lesions.</p><p><strong>Methods: </strong>Data was gathered from patients who were referred to our clinic for an oral mucosal biopsy by one oral medicine specialist. Clinical description, differential diagnoses, and final histopathologic diagnoses were retrospectively extracted from patient records. The lesion description was inputted into ChatGPT version 4.0 under a uniform script to generate three differential diagnoses. ChatGPT and an oral medicine specialist's differential diagnosis were compared to the final histopathologic diagnosis.</p><p><strong>Results: </strong>100 oral soft tissue lesions were evaluated. A statistically significant correlation was found between the ability of the Chatbot and the Specialist to accurately diagnose the cases (P < 0.001). ChatGPT demonstrated remarkable sensitivity for diagnosing urgent cases, as none of the malignant lesions were missed by the chatbot. At the same time, the specificity of the specialist was higher in cases of malignant lesion diagnosis (p < 0.05). The chatbot performance was reliable in two different events (p < 0.01).</p><p><strong>Conclusion: </strong>ChatGPT-4 has shown the ability to pinpoint suspicious malignant lesions and suggest an adequate differential diagnosis for soft tissue lesions, in a consistent and repetitive manner.</p><p><strong>Clinical relevance: </strong>This study serves as a primary insight into the role of AI chatbots, as assisting tools in oral medicine and assesses their clinical capabilities.</p>","PeriodicalId":10461,"journal":{"name":"Clinical Oral Investigations","volume":"29 4","pages":"188"},"PeriodicalIF":3.1000,"publicationDate":"2025-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Clinical Oral Investigations","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1007/s00784-025-06268-7","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"DENTISTRY, ORAL SURGERY & MEDICINE","Score":null,"Total":0}
引用次数: 0
Abstract
Objectives: Artificial intelligence (AI) is becoming more popular in medicine. The current study aims to investigate, primarily, if an AI-based chatbot, such as ChatGPT, could be a valid tool for assisting in establishing a differential diagnosis of oral mucosal lesions.
Methods: Data was gathered from patients who were referred to our clinic for an oral mucosal biopsy by one oral medicine specialist. Clinical description, differential diagnoses, and final histopathologic diagnoses were retrospectively extracted from patient records. The lesion description was inputted into ChatGPT version 4.0 under a uniform script to generate three differential diagnoses. ChatGPT and an oral medicine specialist's differential diagnosis were compared to the final histopathologic diagnosis.
Results: 100 oral soft tissue lesions were evaluated. A statistically significant correlation was found between the ability of the Chatbot and the Specialist to accurately diagnose the cases (P < 0.001). ChatGPT demonstrated remarkable sensitivity for diagnosing urgent cases, as none of the malignant lesions were missed by the chatbot. At the same time, the specificity of the specialist was higher in cases of malignant lesion diagnosis (p < 0.05). The chatbot performance was reliable in two different events (p < 0.01).
Conclusion: ChatGPT-4 has shown the ability to pinpoint suspicious malignant lesions and suggest an adequate differential diagnosis for soft tissue lesions, in a consistent and repetitive manner.
Clinical relevance: This study serves as a primary insight into the role of AI chatbots, as assisting tools in oral medicine and assesses their clinical capabilities.
期刊介绍:
The journal Clinical Oral Investigations is a multidisciplinary, international forum for publication of research from all fields of oral medicine. The journal publishes original scientific articles and invited reviews which provide up-to-date results of basic and clinical studies in oral and maxillofacial science and medicine. The aim is to clarify the relevance of new results to modern practice, for an international readership. Coverage includes maxillofacial and oral surgery, prosthetics and restorative dentistry, operative dentistry, endodontics, periodontology, orthodontics, dental materials science, clinical trials, epidemiology, pedodontics, oral implant, preventive dentistiry, oral pathology, oral basic sciences and more.