{"title":"人机交互中手势的同时分割与识别","authors":"Harold Vasquez Chavarria, H. Escalante, L. Sucar","doi":"10.1109/ICAR.2013.6766511","DOIUrl":null,"url":null,"abstract":"Gestures are a natural form of communication between people that is being increasingly used for human-robot interaction. There are many automatic techniques to recognize gestures, however, most of them assume that gestures are already segmented from continuous video, clearly, this is an unrealistic scenario for human-robot interaction. For instance, when commanding a service robot the agent must be aware at any time of the world (e.g., via continuous video) and ready to react when a user gives an order (e.g., using a gesture). In this paper we propose a method for addressing both tasks, segmentation and recognition of gestures, simultaneously. The proposed method is based on a novel video-stream exploration scheme called multi-size dynamic windows. Several windows of different sizes are dynamically created, each window is classified by a Hidden Markov Model (HMM). Predictions are combined via a voting strategy and eventually the endpoint of a gesture is detected (segmentation). At that moment the method recognizes the gesture that has been just performed using a majority vote decision (recognition). The proposed method is intended to command a service robot by capturing information of user movements with a KinectTM sensor. We evaluated experimentally the proposed method with 5 different gestures suitable for commanding a service robot. Experimental results show that up to 82.76% of the gestures are correctly segmented. The corresponding recognition performance was of 89.58 %. We consider that this performance is acceptable for certain human-robot interaction scenarios.","PeriodicalId":437814,"journal":{"name":"2013 16th International Conference on Advanced Robotics (ICAR)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Simultaneous segmentation and recognition of hand gestures for human-robot interaction\",\"authors\":\"Harold Vasquez Chavarria, H. Escalante, L. Sucar\",\"doi\":\"10.1109/ICAR.2013.6766511\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Gestures are a natural form of communication between people that is being increasingly used for human-robot interaction. There are many automatic techniques to recognize gestures, however, most of them assume that gestures are already segmented from continuous video, clearly, this is an unrealistic scenario for human-robot interaction. For instance, when commanding a service robot the agent must be aware at any time of the world (e.g., via continuous video) and ready to react when a user gives an order (e.g., using a gesture). In this paper we propose a method for addressing both tasks, segmentation and recognition of gestures, simultaneously. The proposed method is based on a novel video-stream exploration scheme called multi-size dynamic windows. Several windows of different sizes are dynamically created, each window is classified by a Hidden Markov Model (HMM). Predictions are combined via a voting strategy and eventually the endpoint of a gesture is detected (segmentation). At that moment the method recognizes the gesture that has been just performed using a majority vote decision (recognition). The proposed method is intended to command a service robot by capturing information of user movements with a KinectTM sensor. We evaluated experimentally the proposed method with 5 different gestures suitable for commanding a service robot. Experimental results show that up to 82.76% of the gestures are correctly segmented. The corresponding recognition performance was of 89.58 %. We consider that this performance is acceptable for certain human-robot interaction scenarios.\",\"PeriodicalId\":437814,\"journal\":{\"name\":\"2013 16th International Conference on Advanced Robotics (ICAR)\",\"volume\":\"5 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2013-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2013 16th International Conference on Advanced Robotics (ICAR)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICAR.2013.6766511\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2013 16th International Conference on Advanced Robotics (ICAR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAR.2013.6766511","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Simultaneous segmentation and recognition of hand gestures for human-robot interaction
Gestures are a natural form of communication between people that is being increasingly used for human-robot interaction. There are many automatic techniques to recognize gestures, however, most of them assume that gestures are already segmented from continuous video, clearly, this is an unrealistic scenario for human-robot interaction. For instance, when commanding a service robot the agent must be aware at any time of the world (e.g., via continuous video) and ready to react when a user gives an order (e.g., using a gesture). In this paper we propose a method for addressing both tasks, segmentation and recognition of gestures, simultaneously. The proposed method is based on a novel video-stream exploration scheme called multi-size dynamic windows. Several windows of different sizes are dynamically created, each window is classified by a Hidden Markov Model (HMM). Predictions are combined via a voting strategy and eventually the endpoint of a gesture is detected (segmentation). At that moment the method recognizes the gesture that has been just performed using a majority vote decision (recognition). The proposed method is intended to command a service robot by capturing information of user movements with a KinectTM sensor. We evaluated experimentally the proposed method with 5 different gestures suitable for commanding a service robot. Experimental results show that up to 82.76% of the gestures are correctly segmented. The corresponding recognition performance was of 89.58 %. We consider that this performance is acceptable for certain human-robot interaction scenarios.