{"title":"机器人辅助外科手术中的手势识别研究","authors":"Hoangminh Huynhnguyen, U. Buy","doi":"10.1109/SA51175.2021.9507175","DOIUrl":null,"url":null,"abstract":"Surgical gesture segmentation and recognition are important steps toward human-robot collaboration in robot-assisted surgery. In the human-robot collaboration paradigm, the robot needs to understand the surgeon's gestures to perform its tasks correctly. Therefore, training a computer vision model to segment and classify gestures in a surgery video is a focus in this field of research. In this paper, we propose a 2-phase surgical gesture recognition method and we evaluate empirically the method on JIGSAWS's suturing video dataset. Our method consists of a 3D convolutional neural network to detect the transition between 2 consecutive surgemes and a convolutional long short-term memory model for surgeme classification. To the best of our knowledge, ours is the first study aimed at detecting action transition in a multi-action video and to classify surgemes using an entire video portion rather than classifying individual frames. We also share our source code at https://github.comfiemiar/surgery-gesture-recog","PeriodicalId":117020,"journal":{"name":"2020 2nd International Conference on Societal Automation (SA)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Toward Gesture Recognition in Robot-Assisted Surgical Procedures\",\"authors\":\"Hoangminh Huynhnguyen, U. Buy\",\"doi\":\"10.1109/SA51175.2021.9507175\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Surgical gesture segmentation and recognition are important steps toward human-robot collaboration in robot-assisted surgery. In the human-robot collaboration paradigm, the robot needs to understand the surgeon's gestures to perform its tasks correctly. Therefore, training a computer vision model to segment and classify gestures in a surgery video is a focus in this field of research. In this paper, we propose a 2-phase surgical gesture recognition method and we evaluate empirically the method on JIGSAWS's suturing video dataset. Our method consists of a 3D convolutional neural network to detect the transition between 2 consecutive surgemes and a convolutional long short-term memory model for surgeme classification. To the best of our knowledge, ours is the first study aimed at detecting action transition in a multi-action video and to classify surgemes using an entire video portion rather than classifying individual frames. We also share our source code at https://github.comfiemiar/surgery-gesture-recog\",\"PeriodicalId\":117020,\"journal\":{\"name\":\"2020 2nd International Conference on Societal Automation (SA)\",\"volume\":\"22 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-05-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 2nd International Conference on Societal Automation (SA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SA51175.2021.9507175\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 2nd International Conference on Societal Automation (SA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SA51175.2021.9507175","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Toward Gesture Recognition in Robot-Assisted Surgical Procedures
Surgical gesture segmentation and recognition are important steps toward human-robot collaboration in robot-assisted surgery. In the human-robot collaboration paradigm, the robot needs to understand the surgeon's gestures to perform its tasks correctly. Therefore, training a computer vision model to segment and classify gestures in a surgery video is a focus in this field of research. In this paper, we propose a 2-phase surgical gesture recognition method and we evaluate empirically the method on JIGSAWS's suturing video dataset. Our method consists of a 3D convolutional neural network to detect the transition between 2 consecutive surgemes and a convolutional long short-term memory model for surgeme classification. To the best of our knowledge, ours is the first study aimed at detecting action transition in a multi-action video and to classify surgemes using an entire video portion rather than classifying individual frames. We also share our source code at https://github.comfiemiar/surgery-gesture-recog