[基于计算机视觉的人工智能检测和识别胃癌根治性腹腔镜胃切除术中的器械和器官:一项多中心研究]。

Q3 Medicine
K C Zhang, Z Qiao, L Yang, T Zhang, F L Liu, D C Sun, T Y Xie, L Guo, C R Lu
{"title":"[基于计算机视觉的人工智能检测和识别胃癌根治性腹腔镜胃切除术中的器械和器官:一项多中心研究]。","authors":"K C Zhang, Z Qiao, L Yang, T Zhang, F L Liu, D C Sun, T Y Xie, L Guo, C R Lu","doi":"10.3760/cma.j.cn441530-20240125-00041","DOIUrl":null,"url":null,"abstract":"<p><p><b>Objective:</b> To investigate the feasibility and accuracy of computer vision-based artificial intelligence technology in detecting and recognizing instruments and organs in the scenario of radical laparoscopic gastrectomy for gastric cancer. <b>Methods:</b> Eight complete laparoscopic distal radical gastrectomy surgery videos were collected from four large tertiary hospitals in China (First Medical Center of Chinese PLA General Hospital [three cases], Liaoning Cancer Hospital [two cases], Liyang Branch of Jiangsu Province People's Hospital [two cases], and Fudan University Shanghai Cancer Center [one case]). PR software was used to extract frames every 5-10 seconds and convert them into image frames. To ensure quality, deduplication was performed manually to remove obvious duplication and blurred image frames. After conversion and deduplication, there were 3369 frame images with a resolution of 1,920×1,080 PPI. LabelMe was used for instance segmentation of the images into the following 23 categories: veins, arteries, sutures, needle holders, ultrasonic knives, suction devices, bleeding, colon, forceps, gallbladder, small gauze, Hem-o-lok, Hem-o-lok appliers, electrocautery hooks, small intestine, hepatogastric ligaments, liver, omentum, pancreas, spleen, surgical staplers, stomach, and trocars. The frame images were randomly allocated to training and validation sets in a 9:1 ratio. The YOLOv8 deep learning framework was used for model training and validation. Precision, recall, average precision (AP), and mean average precision (mAP) were used to evaluate detection and recognition accuracy. <b>Results:</b> The training set contained 3032 frame images comprising 30 895 instance segmentation counts across 23 categories. The validation set contained 337 frame images comprising 3407 instance segmentation counts. The YOLOv8m model was used for training. The loss curve of the training set showed a smooth gradual decrease in loss value as the number of iteration calculations increased. In the training set, the AP values of all 23 categories were above 0.90, with a mAP of 0.99, whereas in the validation set, the mAP of the 23 categories was 0.82. As to individual categories, the AP values for ultrasonic knives, needle holders, forceps, gallbladders, small pieces of gauze, and surgical staplers were 0.96, 0.94, 0.91, 0.91, 0.91, and 0.91, respectively. The model successfully inferred and applied to a 5-minutes video segment of laparoscopic gastroenterostomy suturing. <b>Conclusion:</b> The primary finding of this multicenter study is that computer vision can efficiently, accurately, and in real-time detect organs and instruments in various scenarios of radical laparoscopic gastrectomy for gastric cancer.</p>","PeriodicalId":23959,"journal":{"name":"中华胃肠外科杂志","volume":"27 5","pages":"464-470"},"PeriodicalIF":0.0000,"publicationDate":"2024-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"[Computer-vision-based artificial intelligence for detection and recognition of instruments and organs during radical laparoscopic gastrectomy for gastric cancer: a multicenter study].\",\"authors\":\"K C Zhang, Z Qiao, L Yang, T Zhang, F L Liu, D C Sun, T Y Xie, L Guo, C R Lu\",\"doi\":\"10.3760/cma.j.cn441530-20240125-00041\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p><b>Objective:</b> To investigate the feasibility and accuracy of computer vision-based artificial intelligence technology in detecting and recognizing instruments and organs in the scenario of radical laparoscopic gastrectomy for gastric cancer. <b>Methods:</b> Eight complete laparoscopic distal radical gastrectomy surgery videos were collected from four large tertiary hospitals in China (First Medical Center of Chinese PLA General Hospital [three cases], Liaoning Cancer Hospital [two cases], Liyang Branch of Jiangsu Province People's Hospital [two cases], and Fudan University Shanghai Cancer Center [one case]). PR software was used to extract frames every 5-10 seconds and convert them into image frames. To ensure quality, deduplication was performed manually to remove obvious duplication and blurred image frames. After conversion and deduplication, there were 3369 frame images with a resolution of 1,920×1,080 PPI. LabelMe was used for instance segmentation of the images into the following 23 categories: veins, arteries, sutures, needle holders, ultrasonic knives, suction devices, bleeding, colon, forceps, gallbladder, small gauze, Hem-o-lok, Hem-o-lok appliers, electrocautery hooks, small intestine, hepatogastric ligaments, liver, omentum, pancreas, spleen, surgical staplers, stomach, and trocars. The frame images were randomly allocated to training and validation sets in a 9:1 ratio. The YOLOv8 deep learning framework was used for model training and validation. Precision, recall, average precision (AP), and mean average precision (mAP) were used to evaluate detection and recognition accuracy. <b>Results:</b> The training set contained 3032 frame images comprising 30 895 instance segmentation counts across 23 categories. The validation set contained 337 frame images comprising 3407 instance segmentation counts. The YOLOv8m model was used for training. The loss curve of the training set showed a smooth gradual decrease in loss value as the number of iteration calculations increased. In the training set, the AP values of all 23 categories were above 0.90, with a mAP of 0.99, whereas in the validation set, the mAP of the 23 categories was 0.82. As to individual categories, the AP values for ultrasonic knives, needle holders, forceps, gallbladders, small pieces of gauze, and surgical staplers were 0.96, 0.94, 0.91, 0.91, 0.91, and 0.91, respectively. The model successfully inferred and applied to a 5-minutes video segment of laparoscopic gastroenterostomy suturing. <b>Conclusion:</b> The primary finding of this multicenter study is that computer vision can efficiently, accurately, and in real-time detect organs and instruments in various scenarios of radical laparoscopic gastrectomy for gastric cancer.</p>\",\"PeriodicalId\":23959,\"journal\":{\"name\":\"中华胃肠外科杂志\",\"volume\":\"27 5\",\"pages\":\"464-470\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-05-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"中华胃肠外科杂志\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.3760/cma.j.cn441530-20240125-00041\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"Medicine\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"中华胃肠外科杂志","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.3760/cma.j.cn441530-20240125-00041","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Medicine","Score":null,"Total":0}
引用次数: 0

摘要

目的研究基于计算机视觉的人工智能技术在腹腔镜胃癌根治术中检测和识别器械和器官的可行性和准确性。方法从中国四家大型三甲医院(中国人民解放军总医院第一医学中心[3例]、辽宁省肿瘤医院[2例]、江苏省人民医院溧阳分院[2例]和复旦大学上海肿瘤防治中心[1例])收集8个完整的腹腔镜远端胃癌根治术手术视频。使用 PR 软件每 5-10 秒提取一帧,并将其转换为图像帧。为确保质量,人工进行了重复数据删除,以去除明显重复和模糊的图像帧。经过转换和重复数据删除后,共有 3369 帧图像,分辨率为 1,920×1,080 PPI。LabelMe 用于对图像进行实例分割,将其分为以下 23 类:静脉、动脉、缝合线、针座、超声刀、吸引器、出血、结肠、镊子、胆囊、小纱布、Hem-o-lok、Hem-o-lok 接头、电灼钩、小肠、肝胃韧带、肝脏、网膜、胰腺、脾脏、手术订书机、胃和套管。帧图像以 9:1 的比例随机分配到训练集和验证集。模型训练和验证采用 YOLOv8 深度学习框架。精确度、召回率、平均精确度(AP)和平均平均精确度(mAP)用于评估检测和识别的准确性。结果训练集包含 3032 帧图像,包含 23 个类别的 30 895 个实例分割计数。验证集包含 337 幅帧图像,包含 3407 个实例分割计数。训练使用了 YOLOv8m 模型。训练集的损失曲线显示,随着迭代计算次数的增加,损失值逐渐平稳下降。在训练集中,所有 23 个类别的 AP 值都高于 0.90,mAP 为 0.99,而在验证集中,23 个类别的 mAP 为 0.82。就单个类别而言,超声刀、持针器、镊子、胆囊、小块纱布和手术订书机的 AP 值分别为 0.96、0.94、0.91、0.91、0.91 和 0.91。该模型成功推断并应用于一段 5 分钟的腹腔镜胃肠造口术缝合视频。结论这项多中心研究的主要发现是,计算机视觉可以高效、准确、实时地检测胃癌根治性腹腔镜胃切除术各种场景中的器官和器械。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
[Computer-vision-based artificial intelligence for detection and recognition of instruments and organs during radical laparoscopic gastrectomy for gastric cancer: a multicenter study].

Objective: To investigate the feasibility and accuracy of computer vision-based artificial intelligence technology in detecting and recognizing instruments and organs in the scenario of radical laparoscopic gastrectomy for gastric cancer. Methods: Eight complete laparoscopic distal radical gastrectomy surgery videos were collected from four large tertiary hospitals in China (First Medical Center of Chinese PLA General Hospital [three cases], Liaoning Cancer Hospital [two cases], Liyang Branch of Jiangsu Province People's Hospital [two cases], and Fudan University Shanghai Cancer Center [one case]). PR software was used to extract frames every 5-10 seconds and convert them into image frames. To ensure quality, deduplication was performed manually to remove obvious duplication and blurred image frames. After conversion and deduplication, there were 3369 frame images with a resolution of 1,920×1,080 PPI. LabelMe was used for instance segmentation of the images into the following 23 categories: veins, arteries, sutures, needle holders, ultrasonic knives, suction devices, bleeding, colon, forceps, gallbladder, small gauze, Hem-o-lok, Hem-o-lok appliers, electrocautery hooks, small intestine, hepatogastric ligaments, liver, omentum, pancreas, spleen, surgical staplers, stomach, and trocars. The frame images were randomly allocated to training and validation sets in a 9:1 ratio. The YOLOv8 deep learning framework was used for model training and validation. Precision, recall, average precision (AP), and mean average precision (mAP) were used to evaluate detection and recognition accuracy. Results: The training set contained 3032 frame images comprising 30 895 instance segmentation counts across 23 categories. The validation set contained 337 frame images comprising 3407 instance segmentation counts. The YOLOv8m model was used for training. The loss curve of the training set showed a smooth gradual decrease in loss value as the number of iteration calculations increased. In the training set, the AP values of all 23 categories were above 0.90, with a mAP of 0.99, whereas in the validation set, the mAP of the 23 categories was 0.82. As to individual categories, the AP values for ultrasonic knives, needle holders, forceps, gallbladders, small pieces of gauze, and surgical staplers were 0.96, 0.94, 0.91, 0.91, 0.91, and 0.91, respectively. The model successfully inferred and applied to a 5-minutes video segment of laparoscopic gastroenterostomy suturing. Conclusion: The primary finding of this multicenter study is that computer vision can efficiently, accurately, and in real-time detect organs and instruments in various scenarios of radical laparoscopic gastrectomy for gastric cancer.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
中华胃肠外科杂志
中华胃肠外科杂志 Medicine-Medicine (all)
CiteScore
1.00
自引率
0.00%
发文量
6776
期刊介绍:
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信