{"title":"Toward Gesture Recognition in Robot-Assisted Surgical Procedures","authors":"Hoangminh Huynhnguyen, U. Buy","doi":"10.1109/SA51175.2021.9507175","DOIUrl":null,"url":null,"abstract":"Surgical gesture segmentation and recognition are important steps toward human-robot collaboration in robot-assisted surgery. In the human-robot collaboration paradigm, the robot needs to understand the surgeon's gestures to perform its tasks correctly. Therefore, training a computer vision model to segment and classify gestures in a surgery video is a focus in this field of research. In this paper, we propose a 2-phase surgical gesture recognition method and we evaluate empirically the method on JIGSAWS's suturing video dataset. Our method consists of a 3D convolutional neural network to detect the transition between 2 consecutive surgemes and a convolutional long short-term memory model for surgeme classification. To the best of our knowledge, ours is the first study aimed at detecting action transition in a multi-action video and to classify surgemes using an entire video portion rather than classifying individual frames. We also share our source code at https://github.comfiemiar/surgery-gesture-recog","PeriodicalId":117020,"journal":{"name":"2020 2nd International Conference on Societal Automation (SA)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 2nd International Conference on Societal Automation (SA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SA51175.2021.9507175","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
Surgical gesture segmentation and recognition are important steps toward human-robot collaboration in robot-assisted surgery. In the human-robot collaboration paradigm, the robot needs to understand the surgeon's gestures to perform its tasks correctly. Therefore, training a computer vision model to segment and classify gestures in a surgery video is a focus in this field of research. In this paper, we propose a 2-phase surgical gesture recognition method and we evaluate empirically the method on JIGSAWS's suturing video dataset. Our method consists of a 3D convolutional neural network to detect the transition between 2 consecutive surgemes and a convolutional long short-term memory model for surgeme classification. To the best of our knowledge, ours is the first study aimed at detecting action transition in a multi-action video and to classify surgemes using an entire video portion rather than classifying individual frames. We also share our source code at https://github.comfiemiar/surgery-gesture-recog