{"title":"Self-Prompting Polyp Segmentation in Colonoscopy using Hybrid Yolo-SAM 2 Model","authors":"Mobina Mansoori, Sajjad Shahabodini, Jamshid Abouei, Konstantinos N. Plataniotis, Arash Mohammadi","doi":"arxiv-2409.09484","DOIUrl":null,"url":null,"abstract":"Early diagnosis and treatment of polyps during colonoscopy are essential for\nreducing the incidence and mortality of Colorectal Cancer (CRC). However, the\nvariability in polyp characteristics and the presence of artifacts in\ncolonoscopy images and videos pose significant challenges for accurate and\nefficient polyp detection and segmentation. This paper presents a novel\napproach to polyp segmentation by integrating the Segment Anything Model (SAM\n2) with the YOLOv8 model. Our method leverages YOLOv8's bounding box\npredictions to autonomously generate input prompts for SAM 2, thereby reducing\nthe need for manual annotations. We conducted exhaustive tests on five\nbenchmark colonoscopy image datasets and two colonoscopy video datasets,\ndemonstrating that our method exceeds state-of-the-art models in both image and\nvideo segmentation tasks. Notably, our approach achieves high segmentation\naccuracy using only bounding box annotations, significantly reducing annotation\ntime and effort. This advancement holds promise for enhancing the efficiency\nand scalability of polyp detection in clinical settings\nhttps://github.com/sajjad-sh33/YOLO_SAM2.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - EE - Image and Video Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.09484","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Early diagnosis and treatment of polyps during colonoscopy are essential for
reducing the incidence and mortality of Colorectal Cancer (CRC). However, the
variability in polyp characteristics and the presence of artifacts in
colonoscopy images and videos pose significant challenges for accurate and
efficient polyp detection and segmentation. This paper presents a novel
approach to polyp segmentation by integrating the Segment Anything Model (SAM
2) with the YOLOv8 model. Our method leverages YOLOv8's bounding box
predictions to autonomously generate input prompts for SAM 2, thereby reducing
the need for manual annotations. We conducted exhaustive tests on five
benchmark colonoscopy image datasets and two colonoscopy video datasets,
demonstrating that our method exceeds state-of-the-art models in both image and
video segmentation tasks. Notably, our approach achieves high segmentation
accuracy using only bounding box annotations, significantly reducing annotation
time and effort. This advancement holds promise for enhancing the efficiency
and scalability of polyp detection in clinical settings
https://github.com/sajjad-sh33/YOLO_SAM2.