Sahyun Pak, Sung Gon Park, Jeonghyun Park, Hong Rock Choi, Jun Ho Lee, Wonchul Lee, Sung Tae Cho, Young Goo Lee, Hanjong Ahn
{"title":"Application of deep learning for semantic segmentation in robotic prostatectomy: Comparison of convolutional neural networks and visual transformers.","authors":"Sahyun Pak, Sung Gon Park, Jeonghyun Park, Hong Rock Choi, Jun Ho Lee, Wonchul Lee, Sung Tae Cho, Young Goo Lee, Hanjong Ahn","doi":"10.4111/icu.20240159","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>Semantic segmentation is a fundamental part of the surgical application of deep learning. Traditionally, segmentation in vision tasks has been performed using convolutional neural networks (CNNs), but the transformer architecture has recently been introduced and widely investigated. We aimed to investigate the performance of deep learning models in segmentation in robot-assisted radical prostatectomy (RARP) and identify which of the architectures is superior for segmentation in robotic surgery.</p><p><strong>Materials and methods: </strong>Intraoperative images during RARP were obtained. The dataset was randomly split into training and validation data. Segmentation of the surgical instruments, bladder, prostate, vas and seminal vesicle was performed using three CNN models (DeepLabv3, MANet, and U-Net++) and three transformers (SegFormer, BEiT, and DPT), and their performances were analyzed.</p><p><strong>Results: </strong>The overall segmentation performance during RARP varied across different model architectures. For the CNN models, DeepLabV3 achieved a mean Dice score of 0.938, MANet scored 0.944, and U-Net++ reached 0.930. For the transformer architectures, SegFormer attained a mean Dice score of 0.919, BEiT scored 0.916, and DPT achieved 0.940. The performance of CNN models was superior to that of transformer models in segmenting the prostate, vas, and seminal vesicle.</p><p><strong>Conclusions: </strong>Deep learning models provided accurate segmentation of the surgical instruments and anatomical structures observed during RARP. Both CNN and transformer models showed reliable predictions in the segmentation task; however, CNN models may be more suitable than transformer models for organ segmentation and may be more applicable in unusual cases. Further research with large datasets is needed.</p>","PeriodicalId":14522,"journal":{"name":"Investigative and Clinical Urology","volume":"65 6","pages":"551-558"},"PeriodicalIF":2.5000,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11543645/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Investigative and Clinical Urology","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.4111/icu.20240159","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"UROLOGY & NEPHROLOGY","Score":null,"Total":0}
引用次数: 0
Abstract
Purpose: Semantic segmentation is a fundamental part of the surgical application of deep learning. Traditionally, segmentation in vision tasks has been performed using convolutional neural networks (CNNs), but the transformer architecture has recently been introduced and widely investigated. We aimed to investigate the performance of deep learning models in segmentation in robot-assisted radical prostatectomy (RARP) and identify which of the architectures is superior for segmentation in robotic surgery.
Materials and methods: Intraoperative images during RARP were obtained. The dataset was randomly split into training and validation data. Segmentation of the surgical instruments, bladder, prostate, vas and seminal vesicle was performed using three CNN models (DeepLabv3, MANet, and U-Net++) and three transformers (SegFormer, BEiT, and DPT), and their performances were analyzed.
Results: The overall segmentation performance during RARP varied across different model architectures. For the CNN models, DeepLabV3 achieved a mean Dice score of 0.938, MANet scored 0.944, and U-Net++ reached 0.930. For the transformer architectures, SegFormer attained a mean Dice score of 0.919, BEiT scored 0.916, and DPT achieved 0.940. The performance of CNN models was superior to that of transformer models in segmenting the prostate, vas, and seminal vesicle.
Conclusions: Deep learning models provided accurate segmentation of the surgical instruments and anatomical structures observed during RARP. Both CNN and transformer models showed reliable predictions in the segmentation task; however, CNN models may be more suitable than transformer models for organ segmentation and may be more applicable in unusual cases. Further research with large datasets is needed.
期刊介绍:
Investigative and Clinical Urology (Investig Clin Urol, ICUrology) is an international, peer-reviewed, platinum open access journal published bimonthly. ICUrology aims to provide outstanding scientific and clinical research articles, that will advance knowledge and understanding of urological diseases and current therapeutic treatments. ICUrology publishes Original Articles, Rapid Communications, Review Articles, Special Articles, Innovations in Urology, Editorials, and Letters to the Editor, with a focus on the following areas of expertise:
• Precision Medicine in Urology
• Urological Oncology
• Robotics/Laparoscopy
• Endourology/Urolithiasis
• Lower Urinary Tract Dysfunction
• Female Urology
• Sexual Dysfunction/Infertility
• Infection/Inflammation
• Reconstruction/Transplantation
• Geriatric Urology
• Pediatric Urology
• Basic/Translational Research
One of the notable features of ICUrology is the application of multimedia platforms facilitating easy-to-access online video clips of newly developed surgical techniques from the journal''s website, by a QR (quick response) code located in the article, or via YouTube. ICUrology provides current and highly relevant knowledge to a broad audience at the cutting edge of urological research and clinical practice.