Automated classification of oral potentially malignant disorders and oral squamous cell carcinoma using a convolutional neural network framework: a cross-sectional study
Cristina Saldivia-Siracusa , Eduardo Santos Carlos de Souza , Arnaldo Vitor Barros da Silva , Anna Luíza Damaceno Araújo , Caíque Mariano Pedroso , Tarcília Aparecida da Silva , Maria Sissa Pereira Sant'Ana , Felipe Paiva Fonseca , Hélder Antônio Rebelo Pontes , Marcos G. Quiles , Marcio Ajudarte Lopes , Pablo Agustin Vargas , Syed Ali Khurram , Alexander T. Pearson , Mark W. Lingen , Luiz Paulo Kowalski , Keith D. Hunter , André Carlos Ponce de Leon Ferreira de Carvalho , Alan Roger Santos-Silva
{"title":"Automated classification of oral potentially malignant disorders and oral squamous cell carcinoma using a convolutional neural network framework: a cross-sectional study","authors":"Cristina Saldivia-Siracusa , Eduardo Santos Carlos de Souza , Arnaldo Vitor Barros da Silva , Anna Luíza Damaceno Araújo , Caíque Mariano Pedroso , Tarcília Aparecida da Silva , Maria Sissa Pereira Sant'Ana , Felipe Paiva Fonseca , Hélder Antônio Rebelo Pontes , Marcos G. Quiles , Marcio Ajudarte Lopes , Pablo Agustin Vargas , Syed Ali Khurram , Alexander T. Pearson , Mark W. Lingen , Luiz Paulo Kowalski , Keith D. Hunter , André Carlos Ponce de Leon Ferreira de Carvalho , Alan Roger Santos-Silva","doi":"10.1016/j.lana.2025.101138","DOIUrl":null,"url":null,"abstract":"<div><h3>Background</h3><div>Artificial Intelligence (AI) models hold promise as useful tools in healthcare practice. We aimed to develop and assess AI models for automatic classification of oral potentially malignant disorders (OPMD) and oral squamous cell carcinoma (OSCC) clinical images through a Deep Learning (DL) approach, and to explore explainability using Gradient-weighted Class Activation Mapping (Grad-CAM).</div></div><div><h3>Methods</h3><div>This study assessed a dataset of 778 clinical images of OPMD and OSCC, divided into training, model optimization, and internal testing subsets with an 8:1:1 proportion. Transfer learning strategies were applied to pre-train 8 convolutional neural networks (CNN). Performance was evaluated by mean accuracy, precision, recall, specificity, F1-score and area under the receiver operating characteristic (AUROC) values. Grad-CAM qualitative appraisal was performed to assess explainability.</div></div><div><h3>Findings</h3><div>ConvNeXt and MobileNet CNNs showed the best performance. Transfer learning strategies enhanced performance for both algorithms, and the greatest model achieved mean accuracy, precision, recall, F1-score and AUROC of 0.799, 0.837, 0.756, 0.794 and 0.863 during internal testing, respectively. MobileNet displayed the lowest computational cost. Grad-CAM analysis demonstrated discrepancies between the best-performing model and the highest explainability model.</div></div><div><h3>Interpretation</h3><div>ConvNeXt and MobileNet DL models accurately distinguished OSCC from OPMD in clinical photographs taken with different types of image-capture devices. Grad-CAM proved to be an outstanding tool to improve performance interpretation. Obtained results suggest that the adoption of DL models in healthcare could aid in diagnostic assistance and decision-making during clinical practice.</div></div><div><h3>Funding</h3><div>This work was supported by <span>FAPESP</span> (2022/13069-8, 2022/07276-0, 2021/14585-7 and 2024/20694-1), <span>CAPES</span>, <span>CNPq</span> (307604/2023-3) and <span>FAPEMIG</span>.</div></div>","PeriodicalId":29783,"journal":{"name":"Lancet Regional Health-Americas","volume":"47 ","pages":"Article 101138"},"PeriodicalIF":7.0000,"publicationDate":"2025-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Lancet Regional Health-Americas","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2667193X25001486","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"HEALTH CARE SCIENCES & SERVICES","Score":null,"Total":0}
引用次数: 0
Abstract
Background
Artificial Intelligence (AI) models hold promise as useful tools in healthcare practice. We aimed to develop and assess AI models for automatic classification of oral potentially malignant disorders (OPMD) and oral squamous cell carcinoma (OSCC) clinical images through a Deep Learning (DL) approach, and to explore explainability using Gradient-weighted Class Activation Mapping (Grad-CAM).
Methods
This study assessed a dataset of 778 clinical images of OPMD and OSCC, divided into training, model optimization, and internal testing subsets with an 8:1:1 proportion. Transfer learning strategies were applied to pre-train 8 convolutional neural networks (CNN). Performance was evaluated by mean accuracy, precision, recall, specificity, F1-score and area under the receiver operating characteristic (AUROC) values. Grad-CAM qualitative appraisal was performed to assess explainability.
Findings
ConvNeXt and MobileNet CNNs showed the best performance. Transfer learning strategies enhanced performance for both algorithms, and the greatest model achieved mean accuracy, precision, recall, F1-score and AUROC of 0.799, 0.837, 0.756, 0.794 and 0.863 during internal testing, respectively. MobileNet displayed the lowest computational cost. Grad-CAM analysis demonstrated discrepancies between the best-performing model and the highest explainability model.
Interpretation
ConvNeXt and MobileNet DL models accurately distinguished OSCC from OPMD in clinical photographs taken with different types of image-capture devices. Grad-CAM proved to be an outstanding tool to improve performance interpretation. Obtained results suggest that the adoption of DL models in healthcare could aid in diagnostic assistance and decision-making during clinical practice.
Funding
This work was supported by FAPESP (2022/13069-8, 2022/07276-0, 2021/14585-7 and 2024/20694-1), CAPES, CNPq (307604/2023-3) and FAPEMIG.
期刊介绍:
The Lancet Regional Health – Americas, an open-access journal, contributes to The Lancet's global initiative by focusing on health-care quality and access in the Americas. It aims to advance clinical practice and health policy in the region, promoting better health outcomes. The journal publishes high-quality original research advocating change or shedding light on clinical practice and health policy. It welcomes submissions on various regional health topics, including infectious diseases, non-communicable diseases, child and adolescent health, maternal and reproductive health, emergency care, health policy, and health equity.