{"title":"Recognition of human-human interaction using CWDTW","authors":"T. Subetha, S. Chitrakala","doi":"10.1109/ICCPCT.2016.7530365","DOIUrl":null,"url":null,"abstract":"Understanding the activities of human is a challenging task in Computer Vision. Identifying the activities of human from videos and predicting their activity class label is the key functionality of Human Activity Recognition system. In general the major issues of Human activity recognition system is to identify the activities with or without the concurrent movement of body parts, occlusion, incremental learning etc. Among these issues the major difficulty lies in detecting the activity of human performing the interactions with or without the concurrent movement of their body parts. This paper aims in resolving the aforementioned problem. Here, the frames are extracted from the videos using the conventional frame extraction techniques. A pixel-based Local Binary Similarity Pattern background subtraction algorithm is used to detect the foreground from the extracted frames. The features are extracted from the detected foreground using Histogram of oriented Gradients and pyramidal feature extraction technique. A 20-point Microsoft human kinematic model is constructed using the set of features present in the frame and supervised temporal-stochastic neighbor embedding is applied to transform a high dimensional data to a low dimensional data. K-means clustering is then applied to produce a bag of key poses. The classifier Constrained Weighted Dynamic Time Warping(CWDTW) is used for the final generation of activity class label. Experimental results show the higher recognition rate achieved for various interactions with the benchmarking datasets such as Kinect Interaction dataset and Gaming dataset.","PeriodicalId":431894,"journal":{"name":"2016 International Conference on Circuit, Power and Computing Technologies (ICCPCT)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2016-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 International Conference on Circuit, Power and Computing Technologies (ICCPCT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCPCT.2016.7530365","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
Understanding the activities of human is a challenging task in Computer Vision. Identifying the activities of human from videos and predicting their activity class label is the key functionality of Human Activity Recognition system. In general the major issues of Human activity recognition system is to identify the activities with or without the concurrent movement of body parts, occlusion, incremental learning etc. Among these issues the major difficulty lies in detecting the activity of human performing the interactions with or without the concurrent movement of their body parts. This paper aims in resolving the aforementioned problem. Here, the frames are extracted from the videos using the conventional frame extraction techniques. A pixel-based Local Binary Similarity Pattern background subtraction algorithm is used to detect the foreground from the extracted frames. The features are extracted from the detected foreground using Histogram of oriented Gradients and pyramidal feature extraction technique. A 20-point Microsoft human kinematic model is constructed using the set of features present in the frame and supervised temporal-stochastic neighbor embedding is applied to transform a high dimensional data to a low dimensional data. K-means clustering is then applied to produce a bag of key poses. The classifier Constrained Weighted Dynamic Time Warping(CWDTW) is used for the final generation of activity class label. Experimental results show the higher recognition rate achieved for various interactions with the benchmarking datasets such as Kinect Interaction dataset and Gaming dataset.