{"title":"VSGNet:基于视觉显著性的皮肤病变分割网络","authors":"Zhefei Cai , Yingle Fan , Tao Fang , Wei Wu","doi":"10.1016/j.eswa.2025.128635","DOIUrl":null,"url":null,"abstract":"<div><div>The accuracy of skin lesion segmentation is of great significance for the subsequent clinical diagnosis. In order to improve the segmentation accuracy, some pioneering works tried to embed multiple complex modules, or used the huge Transformer framework, but due to the limitation of computing resources, these type of large models were not suitable for the actual clinical environment. To address the coexistence challenges of precision and lightweight, we propose a visual saliency guided network (VSGNet) for skin lesion segmentation, which generates saliency images of skin lesions through the efficient attention mechanism of biological vision, and guides the network to quickly locate the target area, so as to solve the localization difficulties in the skin lesion segmentation tasks. VSGNet includes three parts: Color Constancy module, Saliency Detection module and Ultra Lightweight Multi-level Interconnection Network (ULMI-Net). Specially, ULMI-Net uses a U-shaped structure network as the skeleton, including the Adaptive Split Channel Attention (ASCA) module that simulates the parallel mechanism of biological vision dual pathway, and the Channel-Spatial Parallel Attention (CSPA) module inspired by the multi-level interconnection structure of visual cortices. Through these modules, ULMI-Net can balance the efficient extraction and multi-scale fusion of global and local features, and try to achieve the excellent segmentation results at the lowest cost of parameters and computational complexity. To validate the effectiveness and robustness of the proposed VSGNet on three publicly available skin lesion segmentation datasets (ISIC2017, ISIC2018 and PH2 datasets). The experimental results show that compared to other state-of-the-art methods, VSGNet improves the Dice and mIoU metrics by 1.84 % and 3.34 %, respectively, and with a 196 × and 106 × reduction in the number of parameters and computational complexity. This paper constructs the VSGNet integrating the biological vision mechanism and the artificial intelligence algorithm, providing a new idea for the construction of deep learning models guided by the biological vision, promoting the development of biomimetic computational vision as well as.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"292 ","pages":"Article 128635"},"PeriodicalIF":7.5000,"publicationDate":"2025-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"VSGNet: visual saliency guided network for skin lesion segmentation\",\"authors\":\"Zhefei Cai , Yingle Fan , Tao Fang , Wei Wu\",\"doi\":\"10.1016/j.eswa.2025.128635\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>The accuracy of skin lesion segmentation is of great significance for the subsequent clinical diagnosis. In order to improve the segmentation accuracy, some pioneering works tried to embed multiple complex modules, or used the huge Transformer framework, but due to the limitation of computing resources, these type of large models were not suitable for the actual clinical environment. To address the coexistence challenges of precision and lightweight, we propose a visual saliency guided network (VSGNet) for skin lesion segmentation, which generates saliency images of skin lesions through the efficient attention mechanism of biological vision, and guides the network to quickly locate the target area, so as to solve the localization difficulties in the skin lesion segmentation tasks. VSGNet includes three parts: Color Constancy module, Saliency Detection module and Ultra Lightweight Multi-level Interconnection Network (ULMI-Net). Specially, ULMI-Net uses a U-shaped structure network as the skeleton, including the Adaptive Split Channel Attention (ASCA) module that simulates the parallel mechanism of biological vision dual pathway, and the Channel-Spatial Parallel Attention (CSPA) module inspired by the multi-level interconnection structure of visual cortices. Through these modules, ULMI-Net can balance the efficient extraction and multi-scale fusion of global and local features, and try to achieve the excellent segmentation results at the lowest cost of parameters and computational complexity. To validate the effectiveness and robustness of the proposed VSGNet on three publicly available skin lesion segmentation datasets (ISIC2017, ISIC2018 and PH2 datasets). The experimental results show that compared to other state-of-the-art methods, VSGNet improves the Dice and mIoU metrics by 1.84 % and 3.34 %, respectively, and with a 196 × and 106 × reduction in the number of parameters and computational complexity. This paper constructs the VSGNet integrating the biological vision mechanism and the artificial intelligence algorithm, providing a new idea for the construction of deep learning models guided by the biological vision, promoting the development of biomimetic computational vision as well as.</div></div>\",\"PeriodicalId\":50461,\"journal\":{\"name\":\"Expert Systems with Applications\",\"volume\":\"292 \",\"pages\":\"Article 128635\"},\"PeriodicalIF\":7.5000,\"publicationDate\":\"2025-06-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Expert Systems with Applications\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0957417425022547\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Expert Systems with Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0957417425022547","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
VSGNet: visual saliency guided network for skin lesion segmentation
The accuracy of skin lesion segmentation is of great significance for the subsequent clinical diagnosis. In order to improve the segmentation accuracy, some pioneering works tried to embed multiple complex modules, or used the huge Transformer framework, but due to the limitation of computing resources, these type of large models were not suitable for the actual clinical environment. To address the coexistence challenges of precision and lightweight, we propose a visual saliency guided network (VSGNet) for skin lesion segmentation, which generates saliency images of skin lesions through the efficient attention mechanism of biological vision, and guides the network to quickly locate the target area, so as to solve the localization difficulties in the skin lesion segmentation tasks. VSGNet includes three parts: Color Constancy module, Saliency Detection module and Ultra Lightweight Multi-level Interconnection Network (ULMI-Net). Specially, ULMI-Net uses a U-shaped structure network as the skeleton, including the Adaptive Split Channel Attention (ASCA) module that simulates the parallel mechanism of biological vision dual pathway, and the Channel-Spatial Parallel Attention (CSPA) module inspired by the multi-level interconnection structure of visual cortices. Through these modules, ULMI-Net can balance the efficient extraction and multi-scale fusion of global and local features, and try to achieve the excellent segmentation results at the lowest cost of parameters and computational complexity. To validate the effectiveness and robustness of the proposed VSGNet on three publicly available skin lesion segmentation datasets (ISIC2017, ISIC2018 and PH2 datasets). The experimental results show that compared to other state-of-the-art methods, VSGNet improves the Dice and mIoU metrics by 1.84 % and 3.34 %, respectively, and with a 196 × and 106 × reduction in the number of parameters and computational complexity. This paper constructs the VSGNet integrating the biological vision mechanism and the artificial intelligence algorithm, providing a new idea for the construction of deep learning models guided by the biological vision, promoting the development of biomimetic computational vision as well as.
期刊介绍:
Expert Systems With Applications is an international journal dedicated to the exchange of information on expert and intelligent systems used globally in industry, government, and universities. The journal emphasizes original papers covering the design, development, testing, implementation, and management of these systems, offering practical guidelines. It spans various sectors such as finance, engineering, marketing, law, project management, information management, medicine, and more. The journal also welcomes papers on multi-agent systems, knowledge management, neural networks, knowledge discovery, data mining, and other related areas, excluding applications to military/defense systems.