Lingping Kong , Juan D. Velasquez , Václav Snášel , Millie Pant , Jeng-Shyang Pan , Jana Nowakova
{"title":"通过类别表示和预训练模型融合增强皮肤癌检测","authors":"Lingping Kong , Juan D. Velasquez , Václav Snášel , Millie Pant , Jeng-Shyang Pan , Jana Nowakova","doi":"10.1016/j.inffus.2025.103369","DOIUrl":null,"url":null,"abstract":"<div><div>The use of pre-trained models in medical image classification has gained significant attention due to their ability to handle complex datasets and improve accuracy. However, challenges such as domain-specific customization, interpretability, and computational efficiency remain critical, especially in high-stakes applications such as skin cancer detection. In this paper, we introduce a novel interpretability-assisted fine-tuning framework that leverages category representation to enhance both model accuracy and transparency.</div><div>Using the widely known HAM10000 dataset, which includes seven imbalanced categories of skin cancer, we demonstrate that our method improves the classification accuracy by 2.6% compared to standard pre-trained models. In addition to precision, we also achieve significant improvements in interpretability, with our category representation framework providing more understandable insights into the model’s decision-making process. Key metrics, such as precision and recall, show enhanced performance, particularly for underrepresented skin cancer types such as Melanocytic Nevi (F1 score of 0.94) and Actinic Keratosis (F1 score of 0.66).</div><div>Furthermore, the prediction accuracy of the proposed model of the top-3 reaches 98. 21%, which is highly significant for medical decision making. This observation in interpretability underscores the value of top-<span><math><mi>n</mi></math></span> predictions, especially in challenging cases, to support more informed and accurate decisions. The proposed fusion framework not only enhances predictive accuracy, but also offers an interpretable model output that can assist clinicians in making informed decisions. This makes our approach particularly relevant in medical applications, where both accuracy and transparency are crucial. Our results highlight the potential of fusing pretrained models with category representation techniques to bridge the gap between performance and interpretability in AI-driven healthcare solutions.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"124 ","pages":"Article 103369"},"PeriodicalIF":15.5000,"publicationDate":"2025-06-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Enhancing skin cancer detection through category representation and fusion of pre-trained models\",\"authors\":\"Lingping Kong , Juan D. Velasquez , Václav Snášel , Millie Pant , Jeng-Shyang Pan , Jana Nowakova\",\"doi\":\"10.1016/j.inffus.2025.103369\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>The use of pre-trained models in medical image classification has gained significant attention due to their ability to handle complex datasets and improve accuracy. However, challenges such as domain-specific customization, interpretability, and computational efficiency remain critical, especially in high-stakes applications such as skin cancer detection. In this paper, we introduce a novel interpretability-assisted fine-tuning framework that leverages category representation to enhance both model accuracy and transparency.</div><div>Using the widely known HAM10000 dataset, which includes seven imbalanced categories of skin cancer, we demonstrate that our method improves the classification accuracy by 2.6% compared to standard pre-trained models. In addition to precision, we also achieve significant improvements in interpretability, with our category representation framework providing more understandable insights into the model’s decision-making process. Key metrics, such as precision and recall, show enhanced performance, particularly for underrepresented skin cancer types such as Melanocytic Nevi (F1 score of 0.94) and Actinic Keratosis (F1 score of 0.66).</div><div>Furthermore, the prediction accuracy of the proposed model of the top-3 reaches 98. 21%, which is highly significant for medical decision making. This observation in interpretability underscores the value of top-<span><math><mi>n</mi></math></span> predictions, especially in challenging cases, to support more informed and accurate decisions. The proposed fusion framework not only enhances predictive accuracy, but also offers an interpretable model output that can assist clinicians in making informed decisions. This makes our approach particularly relevant in medical applications, where both accuracy and transparency are crucial. Our results highlight the potential of fusing pretrained models with category representation techniques to bridge the gap between performance and interpretability in AI-driven healthcare solutions.</div></div>\",\"PeriodicalId\":50367,\"journal\":{\"name\":\"Information Fusion\",\"volume\":\"124 \",\"pages\":\"Article 103369\"},\"PeriodicalIF\":15.5000,\"publicationDate\":\"2025-06-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Fusion\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1566253525004427\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525004427","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Enhancing skin cancer detection through category representation and fusion of pre-trained models
The use of pre-trained models in medical image classification has gained significant attention due to their ability to handle complex datasets and improve accuracy. However, challenges such as domain-specific customization, interpretability, and computational efficiency remain critical, especially in high-stakes applications such as skin cancer detection. In this paper, we introduce a novel interpretability-assisted fine-tuning framework that leverages category representation to enhance both model accuracy and transparency.
Using the widely known HAM10000 dataset, which includes seven imbalanced categories of skin cancer, we demonstrate that our method improves the classification accuracy by 2.6% compared to standard pre-trained models. In addition to precision, we also achieve significant improvements in interpretability, with our category representation framework providing more understandable insights into the model’s decision-making process. Key metrics, such as precision and recall, show enhanced performance, particularly for underrepresented skin cancer types such as Melanocytic Nevi (F1 score of 0.94) and Actinic Keratosis (F1 score of 0.66).
Furthermore, the prediction accuracy of the proposed model of the top-3 reaches 98. 21%, which is highly significant for medical decision making. This observation in interpretability underscores the value of top- predictions, especially in challenging cases, to support more informed and accurate decisions. The proposed fusion framework not only enhances predictive accuracy, but also offers an interpretable model output that can assist clinicians in making informed decisions. This makes our approach particularly relevant in medical applications, where both accuracy and transparency are crucial. Our results highlight the potential of fusing pretrained models with category representation techniques to bridge the gap between performance and interpretability in AI-driven healthcare solutions.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.