Israel Elujide, Chunhai Feng, Aref Shiran, Jian Li, Yonghe Liu
{"title":"Location Independent Gesture Recognition Using Channel State Information","authors":"Israel Elujide, Chunhai Feng, Aref Shiran, Jian Li, Yonghe Liu","doi":"10.1109/CCNC49033.2022.9700590","DOIUrl":null,"url":null,"abstract":"Gesture recognition has been the subject of intensive research in recent years owing to its wide applications. Unlike traditional systems, which usually require wearable sensors, many recent works have achieved the desirable gesture recognition performance using wireless channel state information from commercially available WiFi devices. However, existing works generally require training new models for different locations due to the location-dependent nature of channel state information. This paper proposes a location-independent system that can recognize gestures performed in a new location without training a new model. Our approach uses disentanglement that extricates location and other extraneous information from those needed for gesture recognition. The implementation is based on an unsupervised invariance induction framework consisting of feature extraction, a multi-output latent space, gesture recognition, and decoder modules. The key idea in designing this system is to separate gesture-dependent features from location-dependent features. Specifically, the feature extraction module consisting of a long short-term memory network is employed to select representative features; it essentially serves as an encoder to generate the latent space. During the training process, the network learns to cluster features representation for the gesture recognition and decoder by minimizing the total loss of the gesture recognition and decoder modules. We test our system with a dataset collected from various subjects performing four different gestures in multiple locations in seven rooms with different layouts. The results show that our location-independent gesture recognition system can achieve 88.69% accuracy for new locations.","PeriodicalId":269305,"journal":{"name":"2022 IEEE 19th Annual Consumer Communications & Networking Conference (CCNC)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 19th Annual Consumer Communications & Networking Conference (CCNC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CCNC49033.2022.9700590","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Gesture recognition has been the subject of intensive research in recent years owing to its wide applications. Unlike traditional systems, which usually require wearable sensors, many recent works have achieved the desirable gesture recognition performance using wireless channel state information from commercially available WiFi devices. However, existing works generally require training new models for different locations due to the location-dependent nature of channel state information. This paper proposes a location-independent system that can recognize gestures performed in a new location without training a new model. Our approach uses disentanglement that extricates location and other extraneous information from those needed for gesture recognition. The implementation is based on an unsupervised invariance induction framework consisting of feature extraction, a multi-output latent space, gesture recognition, and decoder modules. The key idea in designing this system is to separate gesture-dependent features from location-dependent features. Specifically, the feature extraction module consisting of a long short-term memory network is employed to select representative features; it essentially serves as an encoder to generate the latent space. During the training process, the network learns to cluster features representation for the gesture recognition and decoder by minimizing the total loss of the gesture recognition and decoder modules. We test our system with a dataset collected from various subjects performing four different gestures in multiple locations in seven rooms with different layouts. The results show that our location-independent gesture recognition system can achieve 88.69% accuracy for new locations.