{"title":"基于标记潜狄利克雷分配模型的人体动作识别","authors":"Jiahui Yang, Changhong Chen, Z. Gan, Xiuchang Zhu","doi":"10.1109/WCSP.2013.6677264","DOIUrl":null,"url":null,"abstract":"Recognition of human actions has already been an active area in the computer vision domain and techniques related to action recognition have been applied in plenty of fields such as smart surveillance, motion analysis and virtual reality. In this paper, we propose a new action recognition method which represents human actions as a bag of spatio-temporal words extracted from input video sequences and uses L-LDA (labeled Latent Dirichlet Allocation) model as a classifier. L-LDA is a supervised model extended from LDA which is unsupervised. The L-LDA adds a label layer on the basis of LDA to label the category of the train video sequences, so L-LDA can assign the latent topic variable in the model to the specific action categorization automatically. What's more, due to above characteristic of L-LDA, it can help to estimate the model parameters more reasonably, accurately and fast. We test our method on the KTH and Weizmann human action dataset and the experimental results show that L-LDA is better than its unsupervised counterpart LDA as well as SVMs (support vector machines).","PeriodicalId":342639,"journal":{"name":"2013 International Conference on Wireless Communications and Signal Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Human action recognition using labeled Latent Dirichlet Allocation model\",\"authors\":\"Jiahui Yang, Changhong Chen, Z. Gan, Xiuchang Zhu\",\"doi\":\"10.1109/WCSP.2013.6677264\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Recognition of human actions has already been an active area in the computer vision domain and techniques related to action recognition have been applied in plenty of fields such as smart surveillance, motion analysis and virtual reality. In this paper, we propose a new action recognition method which represents human actions as a bag of spatio-temporal words extracted from input video sequences and uses L-LDA (labeled Latent Dirichlet Allocation) model as a classifier. L-LDA is a supervised model extended from LDA which is unsupervised. The L-LDA adds a label layer on the basis of LDA to label the category of the train video sequences, so L-LDA can assign the latent topic variable in the model to the specific action categorization automatically. What's more, due to above characteristic of L-LDA, it can help to estimate the model parameters more reasonably, accurately and fast. We test our method on the KTH and Weizmann human action dataset and the experimental results show that L-LDA is better than its unsupervised counterpart LDA as well as SVMs (support vector machines).\",\"PeriodicalId\":342639,\"journal\":{\"name\":\"2013 International Conference on Wireless Communications and Signal Processing\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2013-12-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2013 International Conference on Wireless Communications and Signal Processing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/WCSP.2013.6677264\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2013 International Conference on Wireless Communications and Signal Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WCSP.2013.6677264","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Human action recognition using labeled Latent Dirichlet Allocation model
Recognition of human actions has already been an active area in the computer vision domain and techniques related to action recognition have been applied in plenty of fields such as smart surveillance, motion analysis and virtual reality. In this paper, we propose a new action recognition method which represents human actions as a bag of spatio-temporal words extracted from input video sequences and uses L-LDA (labeled Latent Dirichlet Allocation) model as a classifier. L-LDA is a supervised model extended from LDA which is unsupervised. The L-LDA adds a label layer on the basis of LDA to label the category of the train video sequences, so L-LDA can assign the latent topic variable in the model to the specific action categorization automatically. What's more, due to above characteristic of L-LDA, it can help to estimate the model parameters more reasonably, accurately and fast. We test our method on the KTH and Weizmann human action dataset and the experimental results show that L-LDA is better than its unsupervised counterpart LDA as well as SVMs (support vector machines).