{"title":"基于RGBD相机的人体轮廓提取及其动作识别","authors":"Xiaohui Huang, Jun Cheng, Xiaopeng Ji","doi":"10.1109/ICINFA.2016.7832114","DOIUrl":null,"url":null,"abstract":"Spatio-temporal cuboid pyramid (STCP) for action recognition using depth motion sequences [1] is influenced by depth camera error which leads the depth motion sequence (DMS) existing many kinds of noise, especially on the surface. It means that the dimension of DMS is awfully high and the feature for action recognition becomes less apparent. In this paper, we present an effective method to reduce noise, which is to segment foreground. We firstly segment and extract human contour in the color image using convolutional network model. Then, human contour is re-segmented utilizing depth information. Thirdly we project each frame of the segmented depth sequence onto three views. We finally extract features from cuboids and recognize human actions. The proposed approach is evaluated on three public benchmark datasets, i.e., UTKinect-Action Dataset, MSRActionPairs Dataset and 3D Online Action Dataset. Experimental results show that our method achieves state-of-the-art performance.","PeriodicalId":389619,"journal":{"name":"2016 IEEE International Conference on Information and Automation (ICIA)","volume":"114 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":"{\"title\":\"Human contour extraction from RGBD camera for action recognition\",\"authors\":\"Xiaohui Huang, Jun Cheng, Xiaopeng Ji\",\"doi\":\"10.1109/ICINFA.2016.7832114\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Spatio-temporal cuboid pyramid (STCP) for action recognition using depth motion sequences [1] is influenced by depth camera error which leads the depth motion sequence (DMS) existing many kinds of noise, especially on the surface. It means that the dimension of DMS is awfully high and the feature for action recognition becomes less apparent. In this paper, we present an effective method to reduce noise, which is to segment foreground. We firstly segment and extract human contour in the color image using convolutional network model. Then, human contour is re-segmented utilizing depth information. Thirdly we project each frame of the segmented depth sequence onto three views. We finally extract features from cuboids and recognize human actions. The proposed approach is evaluated on three public benchmark datasets, i.e., UTKinect-Action Dataset, MSRActionPairs Dataset and 3D Online Action Dataset. Experimental results show that our method achieves state-of-the-art performance.\",\"PeriodicalId\":389619,\"journal\":{\"name\":\"2016 IEEE International Conference on Information and Automation (ICIA)\",\"volume\":\"114 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2016-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"5\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2016 IEEE International Conference on Information and Automation (ICIA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICINFA.2016.7832114\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 IEEE International Conference on Information and Automation (ICIA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICINFA.2016.7832114","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Human contour extraction from RGBD camera for action recognition
Spatio-temporal cuboid pyramid (STCP) for action recognition using depth motion sequences [1] is influenced by depth camera error which leads the depth motion sequence (DMS) existing many kinds of noise, especially on the surface. It means that the dimension of DMS is awfully high and the feature for action recognition becomes less apparent. In this paper, we present an effective method to reduce noise, which is to segment foreground. We firstly segment and extract human contour in the color image using convolutional network model. Then, human contour is re-segmented utilizing depth information. Thirdly we project each frame of the segmented depth sequence onto three views. We finally extract features from cuboids and recognize human actions. The proposed approach is evaluated on three public benchmark datasets, i.e., UTKinect-Action Dataset, MSRActionPairs Dataset and 3D Online Action Dataset. Experimental results show that our method achieves state-of-the-art performance.