Ziheng Wang, Xi Liu, Conor Perreault, Anthony Jarc
{"title":"Automatic Detection of Out-of-body Frames in Surgical Videos for Privacy Protection Using Self-supervised Learning and Minimal Labels","authors":"Ziheng Wang, Xi Liu, Conor Perreault, Anthony Jarc","doi":"10.1142/s2424905x23500022","DOIUrl":null,"url":null,"abstract":"Endoscopic video recordings are widely used in minimally invasive robot-assisted surgery, but when the endoscope is outside the patient’s body, it can capture irrelevant segments that may contain sensitive information. To address this, we propose a framework that accurately detects out-of-body frames in surgical videos by leveraging self-supervision with minimal data labels. We use a massive amount of unlabeled endoscopic images to learn meaningful representations in a self-supervised manner. Our approach, which involves pre-training on an auxiliary task and fine-tuning with limited supervision, outperforms previous methods for detecting out-of-body frames in surgical videos captured from da Vinci X and Xi surgical systems. The average F1 scores range from [Formula: see text] to [Formula: see text]. Remarkably, using only [Formula: see text] of the training labels, our approach still maintains an average F1 score performance above 97, outperforming fully-supervised methods with [Formula: see text] fewer labels. These results demonstrate the potential of our framework to facilitate the safe handling of surgical video recordings and enhance data privacy protection in minimally invasive surgery.","PeriodicalId":73821,"journal":{"name":"Journal of medical robotics research","volume":"330 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of medical robotics research","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1142/s2424905x23500022","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Endoscopic video recordings are widely used in minimally invasive robot-assisted surgery, but when the endoscope is outside the patient’s body, it can capture irrelevant segments that may contain sensitive information. To address this, we propose a framework that accurately detects out-of-body frames in surgical videos by leveraging self-supervision with minimal data labels. We use a massive amount of unlabeled endoscopic images to learn meaningful representations in a self-supervised manner. Our approach, which involves pre-training on an auxiliary task and fine-tuning with limited supervision, outperforms previous methods for detecting out-of-body frames in surgical videos captured from da Vinci X and Xi surgical systems. The average F1 scores range from [Formula: see text] to [Formula: see text]. Remarkably, using only [Formula: see text] of the training labels, our approach still maintains an average F1 score performance above 97, outperforming fully-supervised methods with [Formula: see text] fewer labels. These results demonstrate the potential of our framework to facilitate the safe handling of surgical video recordings and enhance data privacy protection in minimally invasive surgery.
内窥镜录像广泛应用于微创机器人辅助手术,但当内窥镜在患者体外时,它可能会捕捉到可能包含敏感信息的不相关片段。为了解决这个问题,我们提出了一个框架,通过利用最小数据标签的自我监督来准确检测手术视频中的体外帧。我们使用大量未标记的内窥镜图像以自我监督的方式学习有意义的表示。F1的平均分数从[公式:见文]到[公式:见文]不等。值得注意的是,仅使用[Formula: see text]的训练标签,我们的方法仍然保持了平均F1分数在97以上的表现,优于使用[Formula: see text]较少标签的完全监督方法。这些结果证明了我们的框架在促进手术视频记录的安全处理和加强微创手术数据隐私保护方面的潜力。