Bilin Wang , Changda Lei , Yunbo Guo , Kaicheng Hong , Xiuji Kan , Yifan Ouyang , Junbo Li , Rui Li
{"title":"Task-specific prompting SAM for multi-task gastric cancer diagnosis in endoscopic images","authors":"Bilin Wang , Changda Lei , Yunbo Guo , Kaicheng Hong , Xiuji Kan , Yifan Ouyang , Junbo Li , Rui Li","doi":"10.1016/j.eswa.2025.128329","DOIUrl":null,"url":null,"abstract":"<div><div>Deep learning is increasingly applied in gastroscopic imaging to assist in lesion detection and diagnosis. However, less attention has been given to analyzing critical lesion characteristics, such as differentiation and invasion depth-factors essential for prognosis and treatment planning. The Segment Anything Model (SAM) introduces robust semantic feature extraction capabilities paired with flexible prompting mechanisms, enabling the model to focus on specific regions of interest. However, designing prompts that effectively capture the diverse semantic information needed for distinct tasks remains underexplored. In this paper, we propose Task-Specific Prompt SAM (TSP-SAM), an innovative framework that employs Task-Specific Prompt Generation (TSPG) and Embedding Prompt Fusion (EPF) to capture multi-perspective features for comprehensive classification and segmentation. Specifically, TSP-SAM utilizes SAM’s powerful feature extraction by integrating multi-scale embeddings and distilled prompt information. This allows the model to refine segmentation, staging, differentiation, and infiltration depth classification within a unified framework. We evaluate TSP-SAM on a newly developed gastroscopy dataset comprising over 3600 images with pixel-level and pathology-based annotations. Extensive experiments demonstrate that TSP-SAM consistently outperforms both traditional and advanced single-task models, showcasing its superior capability for joint optimization across multiple tasks.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"289 ","pages":"Article 128329"},"PeriodicalIF":7.5000,"publicationDate":"2025-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Expert Systems with Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0957417425019487","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Deep learning is increasingly applied in gastroscopic imaging to assist in lesion detection and diagnosis. However, less attention has been given to analyzing critical lesion characteristics, such as differentiation and invasion depth-factors essential for prognosis and treatment planning. The Segment Anything Model (SAM) introduces robust semantic feature extraction capabilities paired with flexible prompting mechanisms, enabling the model to focus on specific regions of interest. However, designing prompts that effectively capture the diverse semantic information needed for distinct tasks remains underexplored. In this paper, we propose Task-Specific Prompt SAM (TSP-SAM), an innovative framework that employs Task-Specific Prompt Generation (TSPG) and Embedding Prompt Fusion (EPF) to capture multi-perspective features for comprehensive classification and segmentation. Specifically, TSP-SAM utilizes SAM’s powerful feature extraction by integrating multi-scale embeddings and distilled prompt information. This allows the model to refine segmentation, staging, differentiation, and infiltration depth classification within a unified framework. We evaluate TSP-SAM on a newly developed gastroscopy dataset comprising over 3600 images with pixel-level and pathology-based annotations. Extensive experiments demonstrate that TSP-SAM consistently outperforms both traditional and advanced single-task models, showcasing its superior capability for joint optimization across multiple tasks.
期刊介绍:
Expert Systems With Applications is an international journal dedicated to the exchange of information on expert and intelligent systems used globally in industry, government, and universities. The journal emphasizes original papers covering the design, development, testing, implementation, and management of these systems, offering practical guidelines. It spans various sectors such as finance, engineering, marketing, law, project management, information management, medicine, and more. The journal also welcomes papers on multi-agent systems, knowledge management, neural networks, knowledge discovery, data mining, and other related areas, excluding applications to military/defense systems.