Evaluation of Net Withdrawal Time and Colonoscopy Video Summarization Using Deep Learning Based Automated Temporal Video Segmentation.

Kanggil Park, Ji Young Lee, Ahin Choi, Jeong-Sik Byeon, Namkug Kim
{"title":"Evaluation of Net Withdrawal Time and Colonoscopy Video Summarization Using Deep Learning Based Automated Temporal Video Segmentation.","authors":"Kanggil Park, Ji Young Lee, Ahin Choi, Jeong-Sik Byeon, Namkug Kim","doi":"10.1007/s10278-025-01632-1","DOIUrl":null,"url":null,"abstract":"<p><p>Adequate withdrawal time is crucial in colonoscopy, as it is directly associated with polyp detection rates. However, traditional withdrawal time measurements can be biased by non-observation activities, leading to inaccurate assessments of procedural quality. This study aimed to develop a deep learning (DL) model that accurately measures net withdrawal time by excluding non-observation phases and generates quantitative visual summaries of key procedural events. We developed a DL-based automated temporal video segmentation model trained on 40 full-length colonoscopy videos and 825 cecum clips extracted from 221 colonoscopy procedures. The model classifies four key events: cecum, intervention, outside, and narrow-band imaging (NBI) mode. Using the temporal video segmentation results, we calculated the net withdrawal time and extracted representative images from each segment for video summarization. Model performance was evaluated using four standard temporal video segmentation metrics, and its correlation with endoscopist-recorded times on both internal and external test datasets. In both internal and external tests, the DL model achieved a total F1 score exceeding 93% for temporal video segmentation performance. The net withdrawal time showed a strong correlation with endoscopist-recorded times (internal dataset, r = 0.984, p < 0.000; external dataset, r = 0.971, p < 0.000). Additionally, the model successfully generated representative images, and the endoscopists' visual assessment confirmed that these images provided accurate summaries of key events. Compared to manual review, the proposed model offers a more efficient, standardized and objective approach to assessing procedural quality. This model has the potential to enhance clinical practice and improve quality assurance in colonoscopy.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2025-08-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of imaging informatics in medicine","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s10278-025-01632-1","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Adequate withdrawal time is crucial in colonoscopy, as it is directly associated with polyp detection rates. However, traditional withdrawal time measurements can be biased by non-observation activities, leading to inaccurate assessments of procedural quality. This study aimed to develop a deep learning (DL) model that accurately measures net withdrawal time by excluding non-observation phases and generates quantitative visual summaries of key procedural events. We developed a DL-based automated temporal video segmentation model trained on 40 full-length colonoscopy videos and 825 cecum clips extracted from 221 colonoscopy procedures. The model classifies four key events: cecum, intervention, outside, and narrow-band imaging (NBI) mode. Using the temporal video segmentation results, we calculated the net withdrawal time and extracted representative images from each segment for video summarization. Model performance was evaluated using four standard temporal video segmentation metrics, and its correlation with endoscopist-recorded times on both internal and external test datasets. In both internal and external tests, the DL model achieved a total F1 score exceeding 93% for temporal video segmentation performance. The net withdrawal time showed a strong correlation with endoscopist-recorded times (internal dataset, r = 0.984, p < 0.000; external dataset, r = 0.971, p < 0.000). Additionally, the model successfully generated representative images, and the endoscopists' visual assessment confirmed that these images provided accurate summaries of key events. Compared to manual review, the proposed model offers a more efficient, standardized and objective approach to assessing procedural quality. This model has the potential to enhance clinical practice and improve quality assurance in colonoscopy.

基于深度学习的自动时间视频分割评估净退出时间和结肠镜视频摘要。
在结肠镜检查中,适当的停药时间是至关重要的,因为它直接关系到息肉的检出率。然而,传统的撤离时间测量可能会受到非观察活动的影响,导致对程序质量的评估不准确。本研究旨在开发一种深度学习(DL)模型,该模型通过排除非观察阶段来准确测量净退出时间,并生成关键程序事件的定量视觉摘要。我们开发了一个基于dl的自动时间视频分割模型,该模型训练了从221个结肠镜检查过程中提取的40个全长结肠镜视频和825个盲肠片段。该模型将四个关键事件分类:盲肠、干预、外部和窄带成像(NBI)模式。利用时域视频分割结果,计算净提取时间,并从每个片段中提取有代表性的图像进行视频摘要。使用四种标准时间视频分割指标评估模型的性能,以及其与内窥镜师在内部和外部测试数据集上记录的时间的相关性。在内部和外部测试中,DL模型的时域视频分割性能F1总分均超过93%。净停药时间与内窥镜医师记录的次数有很强的相关性(内部数据集,r = 0.984, p
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信