Haoneng Lin , Jing Zou , Sen Deng , Ka Po Wong , Angelica I. Aviles-Rivero , Yiting Fan , Alex Pui-Wai Lee , Xiaowei Hu , Jing Qin
{"title":"通过完全三维适应分割任何模型的体积医学图像分割","authors":"Haoneng Lin , Jing Zou , Sen Deng , Ka Po Wong , Angelica I. Aviles-Rivero , Yiting Fan , Alex Pui-Wai Lee , Xiaowei Hu , Jing Qin","doi":"10.1016/j.bbe.2024.11.001","DOIUrl":null,"url":null,"abstract":"<div><div>The Segment Anything Model (SAM) exhibits exceptional generalization capabilities in diverse domains, owing to its interactive learning mechanism designed for precise image segmentation. However, applying SAM to out-of-distribution domains, especially in 3D medical image segmentation, poses challenges. Existing methods for adapting 2D segmentation models to 3D medical data treat 3D volumes as a mere stack of 2D slices. The essential inter-slice information, which is pivotal to faithful 3D medical image segmentation tasks, is unfortunately neglected. In this work, we present the 3D Medical SAM-Adapter (3DMedSAM), a pioneer cross-dimensional adaptation, leveraging SAM’s pre-trained knowledge while accommodating the unique characteristics of 3D medical data. Firstly, to bridge the dimensional gap from 2D to 3D, we design a novel module to replace SAM’s patch embedding, ensuring a seamless transition into 3D image processing and recognition. Besides, we incorporate a 3D Adapter while maintaining the majority of pre-training parameters frozen, enriching deep features with abundant 3D spatial information and achieving efficient fine-tuning. Given the diverse scales of anomalies present in medical images, we also devised a multi-scale 3D mask decoder to elevate the network’s proficiency in medical image segmentation. Through various experiments, we showcase the effectiveness of 3DMedSAM in achieving accurate and robust 3D segmentation on both single-target segmentation and multi-organ segmentation tasks, surpassing the limitations of current methods.</div></div>","PeriodicalId":55381,"journal":{"name":"Biocybernetics and Biomedical Engineering","volume":"45 1","pages":"Pages 1-10"},"PeriodicalIF":5.3000,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Volumetric medical image segmentation via fully 3D adaptation of Segment Anything Model\",\"authors\":\"Haoneng Lin , Jing Zou , Sen Deng , Ka Po Wong , Angelica I. Aviles-Rivero , Yiting Fan , Alex Pui-Wai Lee , Xiaowei Hu , Jing Qin\",\"doi\":\"10.1016/j.bbe.2024.11.001\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>The Segment Anything Model (SAM) exhibits exceptional generalization capabilities in diverse domains, owing to its interactive learning mechanism designed for precise image segmentation. However, applying SAM to out-of-distribution domains, especially in 3D medical image segmentation, poses challenges. Existing methods for adapting 2D segmentation models to 3D medical data treat 3D volumes as a mere stack of 2D slices. The essential inter-slice information, which is pivotal to faithful 3D medical image segmentation tasks, is unfortunately neglected. In this work, we present the 3D Medical SAM-Adapter (3DMedSAM), a pioneer cross-dimensional adaptation, leveraging SAM’s pre-trained knowledge while accommodating the unique characteristics of 3D medical data. Firstly, to bridge the dimensional gap from 2D to 3D, we design a novel module to replace SAM’s patch embedding, ensuring a seamless transition into 3D image processing and recognition. Besides, we incorporate a 3D Adapter while maintaining the majority of pre-training parameters frozen, enriching deep features with abundant 3D spatial information and achieving efficient fine-tuning. Given the diverse scales of anomalies present in medical images, we also devised a multi-scale 3D mask decoder to elevate the network’s proficiency in medical image segmentation. Through various experiments, we showcase the effectiveness of 3DMedSAM in achieving accurate and robust 3D segmentation on both single-target segmentation and multi-organ segmentation tasks, surpassing the limitations of current methods.</div></div>\",\"PeriodicalId\":55381,\"journal\":{\"name\":\"Biocybernetics and Biomedical Engineering\",\"volume\":\"45 1\",\"pages\":\"Pages 1-10\"},\"PeriodicalIF\":5.3000,\"publicationDate\":\"2025-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Biocybernetics and Biomedical Engineering\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0208521624000846\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, BIOMEDICAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Biocybernetics and Biomedical Engineering","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0208521624000846","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
Volumetric medical image segmentation via fully 3D adaptation of Segment Anything Model
The Segment Anything Model (SAM) exhibits exceptional generalization capabilities in diverse domains, owing to its interactive learning mechanism designed for precise image segmentation. However, applying SAM to out-of-distribution domains, especially in 3D medical image segmentation, poses challenges. Existing methods for adapting 2D segmentation models to 3D medical data treat 3D volumes as a mere stack of 2D slices. The essential inter-slice information, which is pivotal to faithful 3D medical image segmentation tasks, is unfortunately neglected. In this work, we present the 3D Medical SAM-Adapter (3DMedSAM), a pioneer cross-dimensional adaptation, leveraging SAM’s pre-trained knowledge while accommodating the unique characteristics of 3D medical data. Firstly, to bridge the dimensional gap from 2D to 3D, we design a novel module to replace SAM’s patch embedding, ensuring a seamless transition into 3D image processing and recognition. Besides, we incorporate a 3D Adapter while maintaining the majority of pre-training parameters frozen, enriching deep features with abundant 3D spatial information and achieving efficient fine-tuning. Given the diverse scales of anomalies present in medical images, we also devised a multi-scale 3D mask decoder to elevate the network’s proficiency in medical image segmentation. Through various experiments, we showcase the effectiveness of 3DMedSAM in achieving accurate and robust 3D segmentation on both single-target segmentation and multi-organ segmentation tasks, surpassing the limitations of current methods.
期刊介绍:
Biocybernetics and Biomedical Engineering is a quarterly journal, founded in 1981, devoted to publishing the results of original, innovative and creative research investigations in the field of Biocybernetics and biomedical engineering, which bridges mathematical, physical, chemical and engineering methods and technology to analyse physiological processes in living organisms as well as to develop methods, devices and systems used in biology and medicine, mainly in medical diagnosis, monitoring systems and therapy. The Journal''s mission is to advance scientific discovery into new or improved standards of care, and promotion a wide-ranging exchange between science and its application to humans.