M. Caon, J. Tscherrig, Yong Yue, Omar Abou Khaled, E. Mugellini
{"title":"Extending the interaction area for view-invariant 3D gesture recognition","authors":"M. Caon, J. Tscherrig, Yong Yue, Omar Abou Khaled, E. Mugellini","doi":"10.1109/IPTA.2012.6469542","DOIUrl":null,"url":null,"abstract":"This paper presents a non-intrusive approach for view-invariant hand gesture recognition. In fact, the representation of gestures changes dynamically depending on camera viewpoints. Therefore, the different positions of the user between the training phase and the evaluation phase can severely compromise the recognition process. The proposed approach involves the calibration of two Microsoft Kinect depth cameras to allow the 3D modeling of the dynamic hands movements. The gestures are modeled as 3D trajectories and the classification is based on Hidden Markov Models. The approach is trained on data from one viewpoint and tested on data from other very different viewpoints with an angular variation of 180°. The average recognition rate is always higher than 94%. Since it is similar to the recognition rate when training and testing on gestures from the same viewpoint, hence the approach is indeed view-invariant. Comparing these results with those deriving from the test of a one depth camera approach demonstrates that the adoption of two calibrated cameras is crucial.","PeriodicalId":267290,"journal":{"name":"2012 3rd International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2012-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2012 3rd International Conference on Image Processing Theory, Tools and Applications (IPTA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IPTA.2012.6469542","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
This paper presents a non-intrusive approach for view-invariant hand gesture recognition. In fact, the representation of gestures changes dynamically depending on camera viewpoints. Therefore, the different positions of the user between the training phase and the evaluation phase can severely compromise the recognition process. The proposed approach involves the calibration of two Microsoft Kinect depth cameras to allow the 3D modeling of the dynamic hands movements. The gestures are modeled as 3D trajectories and the classification is based on Hidden Markov Models. The approach is trained on data from one viewpoint and tested on data from other very different viewpoints with an angular variation of 180°. The average recognition rate is always higher than 94%. Since it is similar to the recognition rate when training and testing on gestures from the same viewpoint, hence the approach is indeed view-invariant. Comparing these results with those deriving from the test of a one depth camera approach demonstrates that the adoption of two calibrated cameras is crucial.