{"title":"DeepPatchNet: A deep learning model for enhanced screening and diagnosis of oral cancer","authors":"Idriss Tafala , Fatima-Ezzahraa Ben-Bouazza , Aymane Edder , Oumaima Manchadi , Bassma Jioudi","doi":"10.1016/j.imu.2025.101658","DOIUrl":null,"url":null,"abstract":"<div><div>Oral cancer remains a serious global health challenge, significantly affecting patient survival and quality of life. While convolutional neural networks (CNNs) have historically dominated image classification tasks, recent advances suggest that transformer-based models may offer superior performance—albeit with high data and computational demands. In this study, we present <strong>DeepPatchNet</strong>, a novel deep learning architecture that integrates DeepLabV3+ and ConvMixer to address these limitations. Designed for histopathological image classification, DeepPatchNet provides a lightweight, interpretable, and high-performing solution. We evaluated the model on the NDB-UFES dataset (3763 images) and an independent H&E-stained OSCC dataset (1020 images), benchmarking its performance against state-of-the-art models including Vision Transformers (ViTs)<span><span>[1]</span></span>, <span><span>[2]</span></span>, InceptionResNetV2, VGG19, and ConvNeXt. DeepPatchNet achieved superior performance with 86.71% accuracy, 86.80% precision, 86.71% recall, and an F1 score of 86.75%, outperforming all comparison models. Furthermore, the integration of Gradient-weighted Class Activation Mapping (Grad-CAM) enhances interpretability by visually highlighting diagnostically relevant features, addressing a key barrier to clinical adoption. While our results are promising, further validation in real-world clinical settings is needed. DeepPatchNet shows strong potential as a reliable decision support tool for early oral cancer detection and diagnosis.</div></div>","PeriodicalId":13953,"journal":{"name":"Informatics in Medicine Unlocked","volume":"57 ","pages":"Article 101658"},"PeriodicalIF":0.0000,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Informatics in Medicine Unlocked","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2352914825000462","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"Medicine","Score":null,"Total":0}
引用次数: 0
Abstract
Oral cancer remains a serious global health challenge, significantly affecting patient survival and quality of life. While convolutional neural networks (CNNs) have historically dominated image classification tasks, recent advances suggest that transformer-based models may offer superior performance—albeit with high data and computational demands. In this study, we present DeepPatchNet, a novel deep learning architecture that integrates DeepLabV3+ and ConvMixer to address these limitations. Designed for histopathological image classification, DeepPatchNet provides a lightweight, interpretable, and high-performing solution. We evaluated the model on the NDB-UFES dataset (3763 images) and an independent H&E-stained OSCC dataset (1020 images), benchmarking its performance against state-of-the-art models including Vision Transformers (ViTs)[1], [2], InceptionResNetV2, VGG19, and ConvNeXt. DeepPatchNet achieved superior performance with 86.71% accuracy, 86.80% precision, 86.71% recall, and an F1 score of 86.75%, outperforming all comparison models. Furthermore, the integration of Gradient-weighted Class Activation Mapping (Grad-CAM) enhances interpretability by visually highlighting diagnostically relevant features, addressing a key barrier to clinical adoption. While our results are promising, further validation in real-world clinical settings is needed. DeepPatchNet shows strong potential as a reliable decision support tool for early oral cancer detection and diagnosis.
期刊介绍:
Informatics in Medicine Unlocked (IMU) is an international gold open access journal covering a broad spectrum of topics within medical informatics, including (but not limited to) papers focusing on imaging, pathology, teledermatology, public health, ophthalmological, nursing and translational medicine informatics. The full papers that are published in the journal are accessible to all who visit the website.