E. Gedat, Pascal Fechner, Richard Fiebelkorn, R. Vandenhouten
{"title":"基于隐马尔可夫模型和神经网络的人体动作识别","authors":"E. Gedat, Pascal Fechner, Richard Fiebelkorn, R. Vandenhouten","doi":"10.1109/SISY.2017.8080544","DOIUrl":null,"url":null,"abstract":"A human action recognition method is introduced that detects a set of actions in videos by a temporal expansion with hidden Markov models of a pose detection with an artificial neural network. The method was set-up and tested using eleven actions from the MOCAP motion capture database comprising 3,947 frames. A poses alphabet of fourteen relevant poses was defined to be learned by an artificial neural network. It was trained with the skeletons and a manual pose classification of 370 key frame images from the motions resulting in an accuracy of 83.5 %. Three actions prevalent in the used motions were chosen to be recognized with the interplay of one separate hidden Markov model for each action. From the output of the trained artificial neural network for all 1,891 frames of seven of the eleven motions together with a manual pose and action classification the matrices of the hidden Markov models for the four actions were calculated. A tailored maximum likelihood estimation based on the Viterbi algorithm generated an action proposal for each frame. The resulting overall accuracy was 83.2 % precision and 83.7 % recall for recognition of the actions.","PeriodicalId":352891,"journal":{"name":"2017 IEEE 15th International Symposium on Intelligent Systems and Informatics (SISY)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Human action recognition with hidden Markov models and neural network derived poses\",\"authors\":\"E. Gedat, Pascal Fechner, Richard Fiebelkorn, R. Vandenhouten\",\"doi\":\"10.1109/SISY.2017.8080544\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"A human action recognition method is introduced that detects a set of actions in videos by a temporal expansion with hidden Markov models of a pose detection with an artificial neural network. The method was set-up and tested using eleven actions from the MOCAP motion capture database comprising 3,947 frames. A poses alphabet of fourteen relevant poses was defined to be learned by an artificial neural network. It was trained with the skeletons and a manual pose classification of 370 key frame images from the motions resulting in an accuracy of 83.5 %. Three actions prevalent in the used motions were chosen to be recognized with the interplay of one separate hidden Markov model for each action. From the output of the trained artificial neural network for all 1,891 frames of seven of the eleven motions together with a manual pose and action classification the matrices of the hidden Markov models for the four actions were calculated. A tailored maximum likelihood estimation based on the Viterbi algorithm generated an action proposal for each frame. The resulting overall accuracy was 83.2 % precision and 83.7 % recall for recognition of the actions.\",\"PeriodicalId\":352891,\"journal\":{\"name\":\"2017 IEEE 15th International Symposium on Intelligent Systems and Informatics (SISY)\",\"volume\":\"25 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2017-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2017 IEEE 15th International Symposium on Intelligent Systems and Informatics (SISY)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SISY.2017.8080544\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 IEEE 15th International Symposium on Intelligent Systems and Informatics (SISY)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SISY.2017.8080544","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Human action recognition with hidden Markov models and neural network derived poses
A human action recognition method is introduced that detects a set of actions in videos by a temporal expansion with hidden Markov models of a pose detection with an artificial neural network. The method was set-up and tested using eleven actions from the MOCAP motion capture database comprising 3,947 frames. A poses alphabet of fourteen relevant poses was defined to be learned by an artificial neural network. It was trained with the skeletons and a manual pose classification of 370 key frame images from the motions resulting in an accuracy of 83.5 %. Three actions prevalent in the used motions were chosen to be recognized with the interplay of one separate hidden Markov model for each action. From the output of the trained artificial neural network for all 1,891 frames of seven of the eleven motions together with a manual pose and action classification the matrices of the hidden Markov models for the four actions were calculated. A tailored maximum likelihood estimation based on the Viterbi algorithm generated an action proposal for each frame. The resulting overall accuracy was 83.2 % precision and 83.7 % recall for recognition of the actions.