Dinh-Tan Pham, Van-Nam Hoang, Viet-Duc Le, Tien Nguyen, Thanh-Hai Tran, Hai Vu, Van-Hung Le, Thi-Lan Le
{"title":"无人机基于骨架动作识别的深度学习模型","authors":"Dinh-Tan Pham, Van-Nam Hoang, Viet-Duc Le, Tien Nguyen, Thanh-Hai Tran, Hai Vu, Van-Hung Le, Thi-Lan Le","doi":"10.1109/ICCE55644.2022.9852103","DOIUrl":null,"url":null,"abstract":"Human action recognition (HAR) is an important task for UAVs for instant decision-making from captured videos. HAR for UAVs is a challenging task due to the UAVs’ motion, attitudes, and view changes during flight. Moreover, UAVs’ video sequences may suffer from blurs and low resolution. All these issues cause difficulty in HAR for UAVs, necessitating the quest for the HAR method that considers UAV data characteristics. In this paper, we revisit some state-of-the-art deep learning methods and evaluate their performance on the UAV-Human dataset- the largest public UAV dataset up to now. Based on the evaluation, we propose a new framework that combines AAGCN and MS-G3D through a Feature Fusion module for data pre-processing in all streams. Experimental results show that our proposed method outperforms state-of-the-art methods on the UAV-Human dataset.","PeriodicalId":388547,"journal":{"name":"2022 IEEE Ninth International Conference on Communications and Electronics (ICCE)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Deep Learning Models for Skeleton-Based Action Recognition for UAVs\",\"authors\":\"Dinh-Tan Pham, Van-Nam Hoang, Viet-Duc Le, Tien Nguyen, Thanh-Hai Tran, Hai Vu, Van-Hung Le, Thi-Lan Le\",\"doi\":\"10.1109/ICCE55644.2022.9852103\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Human action recognition (HAR) is an important task for UAVs for instant decision-making from captured videos. HAR for UAVs is a challenging task due to the UAVs’ motion, attitudes, and view changes during flight. Moreover, UAVs’ video sequences may suffer from blurs and low resolution. All these issues cause difficulty in HAR for UAVs, necessitating the quest for the HAR method that considers UAV data characteristics. In this paper, we revisit some state-of-the-art deep learning methods and evaluate their performance on the UAV-Human dataset- the largest public UAV dataset up to now. Based on the evaluation, we propose a new framework that combines AAGCN and MS-G3D through a Feature Fusion module for data pre-processing in all streams. Experimental results show that our proposed method outperforms state-of-the-art methods on the UAV-Human dataset.\",\"PeriodicalId\":388547,\"journal\":{\"name\":\"2022 IEEE Ninth International Conference on Communications and Electronics (ICCE)\",\"volume\":\"28 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-07-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE Ninth International Conference on Communications and Electronics (ICCE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCE55644.2022.9852103\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE Ninth International Conference on Communications and Electronics (ICCE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCE55644.2022.9852103","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Deep Learning Models for Skeleton-Based Action Recognition for UAVs
Human action recognition (HAR) is an important task for UAVs for instant decision-making from captured videos. HAR for UAVs is a challenging task due to the UAVs’ motion, attitudes, and view changes during flight. Moreover, UAVs’ video sequences may suffer from blurs and low resolution. All these issues cause difficulty in HAR for UAVs, necessitating the quest for the HAR method that considers UAV data characteristics. In this paper, we revisit some state-of-the-art deep learning methods and evaluate their performance on the UAV-Human dataset- the largest public UAV dataset up to now. Based on the evaluation, we propose a new framework that combines AAGCN and MS-G3D through a Feature Fusion module for data pre-processing in all streams. Experimental results show that our proposed method outperforms state-of-the-art methods on the UAV-Human dataset.