{"title":"Skeleton-based Multi-Feature Sharing Real-Time Action Recognition Network for Human-Robot Interaction","authors":"Zhiwen Deng, Qing Gao, Xiang Yu, Zhaojie Ju, Junkang Chen","doi":"10.1109/ICARM58088.2023.10218912","DOIUrl":null,"url":null,"abstract":"Human-Robot Interaction (HRI) is one of the directions that deserves to be studied, and it is used in various fields (e.g., emergency rescue, telemedicine, and astronaut assistance). Action-based HRI has been proven practical in many application scenarios (e.g., industrial teleoperation and telemedicine). Among these many applications, speed and accuracy are two common optimization goals for researchers. A skeleton-based multi-feature sharing real-time action recognition network (MSR-Net) has been proposed to ensure its high accuracy and speed. Three pairs of feature inputs are proposed to make the network input more informative, which contain: joint distance-joint distance motion (JD-JDM), angle-angle motion (A-AM), slow motion coordinates-fast motion coordinates (SMC-FMC). To ensure that features are fully extracted while reducing the number of model parameters, one-dimensional convolutional neural networks (1DCNN) are used to build multi-feature sharing two-stage feature extraction networks. As a result, MSR-Net outperforms state-of-the-art models in accuracy on the JHMDB (86.8%) and SHREC datasets (96.8% on coarse class and 93.8% on fine class). Application experiments were carried out on the HRI platform to demonstrate the effectiveness of MSR-Net in HRI. The video demonstration of the experimental results of the HRI application can be accessed on https://youtu.be/NzXrngh7BbQ.","PeriodicalId":220013,"journal":{"name":"2023 International Conference on Advanced Robotics and Mechatronics (ICARM)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 International Conference on Advanced Robotics and Mechatronics (ICARM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICARM58088.2023.10218912","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Human-Robot Interaction (HRI) is one of the directions that deserves to be studied, and it is used in various fields (e.g., emergency rescue, telemedicine, and astronaut assistance). Action-based HRI has been proven practical in many application scenarios (e.g., industrial teleoperation and telemedicine). Among these many applications, speed and accuracy are two common optimization goals for researchers. A skeleton-based multi-feature sharing real-time action recognition network (MSR-Net) has been proposed to ensure its high accuracy and speed. Three pairs of feature inputs are proposed to make the network input more informative, which contain: joint distance-joint distance motion (JD-JDM), angle-angle motion (A-AM), slow motion coordinates-fast motion coordinates (SMC-FMC). To ensure that features are fully extracted while reducing the number of model parameters, one-dimensional convolutional neural networks (1DCNN) are used to build multi-feature sharing two-stage feature extraction networks. As a result, MSR-Net outperforms state-of-the-art models in accuracy on the JHMDB (86.8%) and SHREC datasets (96.8% on coarse class and 93.8% on fine class). Application experiments were carried out on the HRI platform to demonstrate the effectiveness of MSR-Net in HRI. The video demonstration of the experimental results of the HRI application can be accessed on https://youtu.be/NzXrngh7BbQ.