Cheng Zhao , Chunlun Xiao , Feifei Jin , Anqi Zhu , Zhuo Xiang , Yiyao Liu , Lehang Guo , Tianfu Wang , Baiying Lei
{"title":"基于两阶段跨模态融合的标签引导图学习网络用于多标签皮肤病诊断","authors":"Cheng Zhao , Chunlun Xiao , Feifei Jin , Anqi Zhu , Zhuo Xiang , Yiyao Liu , Lehang Guo , Tianfu Wang , Baiying Lei","doi":"10.1016/j.engappai.2025.112025","DOIUrl":null,"url":null,"abstract":"<div><div>The accurate diagnosis of skin diseases relies on combining cortical lesion morphological features from clinical and dermoscopic data with deep subcutaneous lesion characteristics from ultrasound data. However, current multi-modal diagnostic methods mainly emphasize clinical and dermoscopic data analysis, lacking a thorough exploration of subcutaneous tissue features provided by skin ultrasound. Additionally, existing research often overlooks the analysis of relationships between skin labels and categories across different diagnostic tasks, constraining model performance as the number of target tasks grows. This paper proposes a label-guided graph learning network (LGL_Net) based on two-stage cross-modal fusion to achieve accurate diagnosis of skin diseases. Specifically, this paper first establishes a two-stage cross-modal fusion unit (TC_Fusion) to enable the fusion and transmission of morphological features from clinical data and deep lesion features from ultrasound data. Subsequently, a label-guided graph learning unit (LGL_Unit) is constructed to explore the correlations between multi-label data of skin lesions by building a graph convolutional network (GCN) at the label and category levels, thereby improving the accuracy of various skin disease diagnostic tasks. Extensive experiments were conducted on a public dataset (including clinical and dermoscopic data) and a private dataset (including clinical and ultrasound data). The experimental results demonstrate that the proposed method achieves optimal performance on tasks like pathological diagnosis (PG), benign or malignant diagnosis (BM), and 7-Point Checklist scoring (7 PC), offering a new approach for skin disease diagnosis. Our code is available at: <span><span>https://github.com/Zhaocheng1/LGL-Net</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"160 ","pages":"Article 112025"},"PeriodicalIF":8.0000,"publicationDate":"2025-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Label-guided graph learning network via two-stage cross-modal fusion for multi-label skin disease diagnosis\",\"authors\":\"Cheng Zhao , Chunlun Xiao , Feifei Jin , Anqi Zhu , Zhuo Xiang , Yiyao Liu , Lehang Guo , Tianfu Wang , Baiying Lei\",\"doi\":\"10.1016/j.engappai.2025.112025\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>The accurate diagnosis of skin diseases relies on combining cortical lesion morphological features from clinical and dermoscopic data with deep subcutaneous lesion characteristics from ultrasound data. However, current multi-modal diagnostic methods mainly emphasize clinical and dermoscopic data analysis, lacking a thorough exploration of subcutaneous tissue features provided by skin ultrasound. Additionally, existing research often overlooks the analysis of relationships between skin labels and categories across different diagnostic tasks, constraining model performance as the number of target tasks grows. This paper proposes a label-guided graph learning network (LGL_Net) based on two-stage cross-modal fusion to achieve accurate diagnosis of skin diseases. Specifically, this paper first establishes a two-stage cross-modal fusion unit (TC_Fusion) to enable the fusion and transmission of morphological features from clinical data and deep lesion features from ultrasound data. Subsequently, a label-guided graph learning unit (LGL_Unit) is constructed to explore the correlations between multi-label data of skin lesions by building a graph convolutional network (GCN) at the label and category levels, thereby improving the accuracy of various skin disease diagnostic tasks. Extensive experiments were conducted on a public dataset (including clinical and dermoscopic data) and a private dataset (including clinical and ultrasound data). The experimental results demonstrate that the proposed method achieves optimal performance on tasks like pathological diagnosis (PG), benign or malignant diagnosis (BM), and 7-Point Checklist scoring (7 PC), offering a new approach for skin disease diagnosis. Our code is available at: <span><span>https://github.com/Zhaocheng1/LGL-Net</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":50523,\"journal\":{\"name\":\"Engineering Applications of Artificial Intelligence\",\"volume\":\"160 \",\"pages\":\"Article 112025\"},\"PeriodicalIF\":8.0000,\"publicationDate\":\"2025-08-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Engineering Applications of Artificial Intelligence\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0952197625020330\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0952197625020330","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
Label-guided graph learning network via two-stage cross-modal fusion for multi-label skin disease diagnosis
The accurate diagnosis of skin diseases relies on combining cortical lesion morphological features from clinical and dermoscopic data with deep subcutaneous lesion characteristics from ultrasound data. However, current multi-modal diagnostic methods mainly emphasize clinical and dermoscopic data analysis, lacking a thorough exploration of subcutaneous tissue features provided by skin ultrasound. Additionally, existing research often overlooks the analysis of relationships between skin labels and categories across different diagnostic tasks, constraining model performance as the number of target tasks grows. This paper proposes a label-guided graph learning network (LGL_Net) based on two-stage cross-modal fusion to achieve accurate diagnosis of skin diseases. Specifically, this paper first establishes a two-stage cross-modal fusion unit (TC_Fusion) to enable the fusion and transmission of morphological features from clinical data and deep lesion features from ultrasound data. Subsequently, a label-guided graph learning unit (LGL_Unit) is constructed to explore the correlations between multi-label data of skin lesions by building a graph convolutional network (GCN) at the label and category levels, thereby improving the accuracy of various skin disease diagnostic tasks. Extensive experiments were conducted on a public dataset (including clinical and dermoscopic data) and a private dataset (including clinical and ultrasound data). The experimental results demonstrate that the proposed method achieves optimal performance on tasks like pathological diagnosis (PG), benign or malignant diagnosis (BM), and 7-Point Checklist scoring (7 PC), offering a new approach for skin disease diagnosis. Our code is available at: https://github.com/Zhaocheng1/LGL-Net.
期刊介绍:
Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.