Ying Chen , Wenjing Cui , Xiaoyan Dong , Shuai Zhou , Zhongqiu Wang
{"title":"SAM2Med3D:利用视频基础模型进行3D乳房MRI分割","authors":"Ying Chen , Wenjing Cui , Xiaoyan Dong , Shuai Zhou , Zhongqiu Wang","doi":"10.1016/j.cag.2025.104341","DOIUrl":null,"url":null,"abstract":"<div><div>Foundation models such as the Segment Anything Model 2 (SAM2) have demonstrated impressive generalization across natural image domains. However, their potential in volumetric medical imaging remains largely underexplored, particularly under limited data conditions. In this paper, we present SAM2Med3D, a novel multi-stage framework that adapts a general-purpose video foundation model for accurate and consistent 3D breast MRI segmentation by treating 3D MRI scan as a sequence of images. Unlike existing image-based approaches (e.g., MedSAM) that require large-scale medical data for fine-tuning, our method combines a lightweight, task-specific segmentation network with a video foundation model, achieving strong performance with only modest training data. To guide the foundation model effectively, we introduce a novel spatial filtering strategy that identifies reliable slices from the initial segmentation to serve as high-quality prompts. Additionally, we propose a confidence-driven fusion mechanism that adaptively integrates coarse and refined predictions across the volume, mitigating segmentation drift and ensuring both local accuracy and global volumetric consistency. We validate SAM2Med3D on two multi-center breast MRI datasets, including both public and self-collected datasets. Experimental results demonstrate that our method outperforms both task-specific segmentation networks and recent foundation-model-based methods, achieving superior accuracy and inter-slice consistency.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"132 ","pages":"Article 104341"},"PeriodicalIF":2.8000,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"SAM2Med3D: Leveraging video foundation models for 3D breast MRI segmentation\",\"authors\":\"Ying Chen , Wenjing Cui , Xiaoyan Dong , Shuai Zhou , Zhongqiu Wang\",\"doi\":\"10.1016/j.cag.2025.104341\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Foundation models such as the Segment Anything Model 2 (SAM2) have demonstrated impressive generalization across natural image domains. However, their potential in volumetric medical imaging remains largely underexplored, particularly under limited data conditions. In this paper, we present SAM2Med3D, a novel multi-stage framework that adapts a general-purpose video foundation model for accurate and consistent 3D breast MRI segmentation by treating 3D MRI scan as a sequence of images. Unlike existing image-based approaches (e.g., MedSAM) that require large-scale medical data for fine-tuning, our method combines a lightweight, task-specific segmentation network with a video foundation model, achieving strong performance with only modest training data. To guide the foundation model effectively, we introduce a novel spatial filtering strategy that identifies reliable slices from the initial segmentation to serve as high-quality prompts. Additionally, we propose a confidence-driven fusion mechanism that adaptively integrates coarse and refined predictions across the volume, mitigating segmentation drift and ensuring both local accuracy and global volumetric consistency. We validate SAM2Med3D on two multi-center breast MRI datasets, including both public and self-collected datasets. Experimental results demonstrate that our method outperforms both task-specific segmentation networks and recent foundation-model-based methods, achieving superior accuracy and inter-slice consistency.</div></div>\",\"PeriodicalId\":50628,\"journal\":{\"name\":\"Computers & Graphics-Uk\",\"volume\":\"132 \",\"pages\":\"Article 104341\"},\"PeriodicalIF\":2.8000,\"publicationDate\":\"2025-08-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computers & Graphics-Uk\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0097849325001827\",\"RegionNum\":4,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, SOFTWARE ENGINEERING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computers & Graphics-Uk","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0097849325001827","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
SAM2Med3D: Leveraging video foundation models for 3D breast MRI segmentation
Foundation models such as the Segment Anything Model 2 (SAM2) have demonstrated impressive generalization across natural image domains. However, their potential in volumetric medical imaging remains largely underexplored, particularly under limited data conditions. In this paper, we present SAM2Med3D, a novel multi-stage framework that adapts a general-purpose video foundation model for accurate and consistent 3D breast MRI segmentation by treating 3D MRI scan as a sequence of images. Unlike existing image-based approaches (e.g., MedSAM) that require large-scale medical data for fine-tuning, our method combines a lightweight, task-specific segmentation network with a video foundation model, achieving strong performance with only modest training data. To guide the foundation model effectively, we introduce a novel spatial filtering strategy that identifies reliable slices from the initial segmentation to serve as high-quality prompts. Additionally, we propose a confidence-driven fusion mechanism that adaptively integrates coarse and refined predictions across the volume, mitigating segmentation drift and ensuring both local accuracy and global volumetric consistency. We validate SAM2Med3D on two multi-center breast MRI datasets, including both public and self-collected datasets. Experimental results demonstrate that our method outperforms both task-specific segmentation networks and recent foundation-model-based methods, achieving superior accuracy and inter-slice consistency.
期刊介绍:
Computers & Graphics is dedicated to disseminate information on research and applications of computer graphics (CG) techniques. The journal encourages articles on:
1. Research and applications of interactive computer graphics. We are particularly interested in novel interaction techniques and applications of CG to problem domains.
2. State-of-the-art papers on late-breaking, cutting-edge research on CG.
3. Information on innovative uses of graphics principles and technologies.
4. Tutorial papers on both teaching CG principles and innovative uses of CG in education.