{"title":"基于混合深度学习模型和Shearlet变换的人体动作识别","authors":"Nemir Al-Azzawi","doi":"10.1109/ICITEE49829.2020.9271687","DOIUrl":null,"url":null,"abstract":"The hybrid deep learning model has become common in all recent studies dealing with machine vision and human action recognition. Most of the accuracy in revealing knowledge of machine vision is in extracting important features, including segmentation of the image. This paper proposes a new model for recognizing human actions from video sequences by integrating repetitive, gated recurrent neural networks across multiple scales with shearlet-based image segmentation extraction. Segmentations are the most critical information to distinguish human action. The feature extraction can impact the complexity of the calculation and the performance of the algorithm. The idea is to increase training robustness and improve segmentation through the use of the shearlet transform. Hence, the video classification based on a recurrent neural network and shearlet transform will work optimally. The proposed approach is evaluated on human activity videos using KTH, UCF-101, and UCF Sports Action datasets. The experimental results showed state-of-the-art performance in comparison to current methods. The average resulting classification accuracy is 95.1% for the KTH datasets. That was the optimal case in our proposed model reached.","PeriodicalId":245013,"journal":{"name":"2020 12th International Conference on Information Technology and Electrical Engineering (ICITEE)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Human Action Recognition based on Hybrid Deep Learning Model and Shearlet Transform\",\"authors\":\"Nemir Al-Azzawi\",\"doi\":\"10.1109/ICITEE49829.2020.9271687\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The hybrid deep learning model has become common in all recent studies dealing with machine vision and human action recognition. Most of the accuracy in revealing knowledge of machine vision is in extracting important features, including segmentation of the image. This paper proposes a new model for recognizing human actions from video sequences by integrating repetitive, gated recurrent neural networks across multiple scales with shearlet-based image segmentation extraction. Segmentations are the most critical information to distinguish human action. The feature extraction can impact the complexity of the calculation and the performance of the algorithm. The idea is to increase training robustness and improve segmentation through the use of the shearlet transform. Hence, the video classification based on a recurrent neural network and shearlet transform will work optimally. The proposed approach is evaluated on human activity videos using KTH, UCF-101, and UCF Sports Action datasets. The experimental results showed state-of-the-art performance in comparison to current methods. The average resulting classification accuracy is 95.1% for the KTH datasets. That was the optimal case in our proposed model reached.\",\"PeriodicalId\":245013,\"journal\":{\"name\":\"2020 12th International Conference on Information Technology and Electrical Engineering (ICITEE)\",\"volume\":\"37 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-10-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 12th International Conference on Information Technology and Electrical Engineering (ICITEE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICITEE49829.2020.9271687\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 12th International Conference on Information Technology and Electrical Engineering (ICITEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICITEE49829.2020.9271687","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Human Action Recognition based on Hybrid Deep Learning Model and Shearlet Transform
The hybrid deep learning model has become common in all recent studies dealing with machine vision and human action recognition. Most of the accuracy in revealing knowledge of machine vision is in extracting important features, including segmentation of the image. This paper proposes a new model for recognizing human actions from video sequences by integrating repetitive, gated recurrent neural networks across multiple scales with shearlet-based image segmentation extraction. Segmentations are the most critical information to distinguish human action. The feature extraction can impact the complexity of the calculation and the performance of the algorithm. The idea is to increase training robustness and improve segmentation through the use of the shearlet transform. Hence, the video classification based on a recurrent neural network and shearlet transform will work optimally. The proposed approach is evaluated on human activity videos using KTH, UCF-101, and UCF Sports Action datasets. The experimental results showed state-of-the-art performance in comparison to current methods. The average resulting classification accuracy is 95.1% for the KTH datasets. That was the optimal case in our proposed model reached.