Non-invasive Primary Screening of Oral Lesions into Binary and Multi Class using Convolutional Neural Network, Stratified K-fold Validation and Transfer Learning
{"title":"Non-invasive Primary Screening of Oral Lesions into Binary and Multi Class using Convolutional Neural Network, Stratified K-fold Validation and Transfer Learning","authors":"Rinkal Shah, Jyoti Pareek","doi":"10.17485/ijst/v17i7.2670","DOIUrl":null,"url":null,"abstract":"Objectives: To develop a deep learning method using camera images that can effectively detect the preliminary phase of oral cancer, which has a high rate of morbidity and mortality and is a significant public health concern. If left untreated, it can result in severe deformities and negatively affect the patient's quality of life, both physically and mentally. Early detection is crucial owing to the rapid spread of the disease, where biopsy is the only option left. Therefore, it is essential to identify malignancies swiftly to prevent disease progression non-invasively. Methods: Three different scenarios are used in this study to analyze samples: CNN architecture, stratified K-fold validation, and transfer learning. For automated disease identification on binary datasets (normal vs. ulcer) and multiclass datasets (normal vs. ulcer vs. Leukoplakia), camera images are pre-processed with data augmentation. As a feature extractor in the model, transfer learning is used with pre-defined networks such as VGG19, InceptionNET, EfficientNET, and MobileNET weights. Findings: Using the proposed CNN architecture, the F1 score for image classification was 78% and 74% for photos showing hygienic mouths or ulcers, and 83%, 87%, and 84% for images showing normal mouths, ulcers, and leukoplakia. Using stratified 3-fold validation, the results were improved to 97%, and an EfficientNET achieves the highest results in a binary F1 score of 98% and a classification with multiple classes F1 scores of 98%, 87%, and 91%, respectively. Novelty: Previous studies have mostly concentrated on differentiating oral potentially malignant diseases (OPMD) from oral squamous cell carcinoma (OSCC) or on discriminating between cancerous and non-cancerous tissues. The objective is to diagnose patients with non-invasive procedures to classify ulcers, healthy mouths, or precancerous type \"Leukoplakia\" without requiring them to visit a doctor. Keywords: CNN, Transfer Learning, Oral Cancer, Ulcer, Leukoplakia, Stratified K-fold validation","PeriodicalId":508200,"journal":{"name":"Indian Journal Of Science And Technology","volume":"36 3","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Indian Journal Of Science And Technology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.17485/ijst/v17i7.2670","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Objectives: To develop a deep learning method using camera images that can effectively detect the preliminary phase of oral cancer, which has a high rate of morbidity and mortality and is a significant public health concern. If left untreated, it can result in severe deformities and negatively affect the patient's quality of life, both physically and mentally. Early detection is crucial owing to the rapid spread of the disease, where biopsy is the only option left. Therefore, it is essential to identify malignancies swiftly to prevent disease progression non-invasively. Methods: Three different scenarios are used in this study to analyze samples: CNN architecture, stratified K-fold validation, and transfer learning. For automated disease identification on binary datasets (normal vs. ulcer) and multiclass datasets (normal vs. ulcer vs. Leukoplakia), camera images are pre-processed with data augmentation. As a feature extractor in the model, transfer learning is used with pre-defined networks such as VGG19, InceptionNET, EfficientNET, and MobileNET weights. Findings: Using the proposed CNN architecture, the F1 score for image classification was 78% and 74% for photos showing hygienic mouths or ulcers, and 83%, 87%, and 84% for images showing normal mouths, ulcers, and leukoplakia. Using stratified 3-fold validation, the results were improved to 97%, and an EfficientNET achieves the highest results in a binary F1 score of 98% and a classification with multiple classes F1 scores of 98%, 87%, and 91%, respectively. Novelty: Previous studies have mostly concentrated on differentiating oral potentially malignant diseases (OPMD) from oral squamous cell carcinoma (OSCC) or on discriminating between cancerous and non-cancerous tissues. The objective is to diagnose patients with non-invasive procedures to classify ulcers, healthy mouths, or precancerous type "Leukoplakia" without requiring them to visit a doctor. Keywords: CNN, Transfer Learning, Oral Cancer, Ulcer, Leukoplakia, Stratified K-fold validation
目的开发一种利用摄像头图像的深度学习方法,该方法可有效检测口腔癌的初期阶段,口腔癌的发病率和死亡率都很高,是一个重大的公共卫生问题。如果不及时治疗,口腔癌会导致严重畸形,并对患者的身心生活质量造成负面影响。由于疾病传播迅速,活检是唯一的选择,因此早期发现至关重要。因此,必须迅速识别恶性肿瘤,以非侵入性的方式防止疾病恶化。方法:本研究采用了三种不同的方案来分析样本:CNN 架构、分层 K 折验证和迁移学习。为了在二元数据集(正常 vs. 溃疡)和多类数据集(正常 vs. 溃疡 vs. 白斑病)上自动识别疾病,对摄像头图像进行了数据增强预处理。作为模型中的特征提取器,迁移学习使用了预先定义的网络,如 VGG19、InceptionNET、EfficientNET 和 MobileNET 权重。研究结果使用提出的 CNN 架构,对显示卫生口腔或溃疡的照片进行图像分类的 F1 分数分别为 78% 和 74%,对显示正常口腔、溃疡和白斑病的图像进行分类的 F1 分数分别为 83%、87% 和 84%。通过分层 3 倍验证,结果提高到 97%,EfficientNET 的二元 F1 得分达到 98%,多类分类 F1 得分分别为 98%、87% 和 91%,取得了最高成绩。新颖性:以往的研究大多集中于区分口腔潜在恶性疾病(OPMD)和口腔鳞状细胞癌(OSCC),或区分癌组织和非癌组织。本研究的目的是通过非侵入性程序对患者进行诊断,对溃疡、健康口腔或癌前病变类型 "白斑病 "进行分类,而无需患者就医。关键词CNN、迁移学习、口腔癌、溃疡、白斑病、分层 K 倍验证