{"title":"ViWise:融合视觉和无线传感数据用于轨迹关系识别","authors":"Fang-Jing Wu, Sheng-Wun Lai, Sok-Ian Sou","doi":"10.1145/3614441","DOIUrl":null,"url":null,"abstract":"People usually form a social structure (e.g., a leader-follower, companion, or independent group) for better interactions among them and thus share similar perceptions of visible scenes and invisible wireless signals encountered while moving. Many mobility-driven applications have paid much attention to recognizing trajectory relationships among people. This work models visual and wireless data to quantify the trajectory similarity between a pair of users. We design a visual and wireless sensor fusion system, called ViWise, which incorporates the first-person video frames collected by a wearable visual device and the wireless packets broadcast by a personal mobile device for recognizing finer-grained trajectory relationships within a mobility group. When people take similar trajectories, they usually share similar visual scenes. Their wireless packets observed by ambient wireless base stations (called wireless scanners in this work) usually contain similar patterns. We model the visual characteristics of physical objects seen by a user from two perspectives: micro-scale image structure with pixel-wise features and macro-scale semantic context. On the other hand, we model characteristics of wireless packets based on the encountered wireless scanners along the user’s trajectory. Given two users’ trajectories, their trajectory characteristics behind the visible video frames and invisible wireless packets are fused together to compute the visual-wireless data similarity that quantifies the correlation between trajectories taken by them. We exploit modeled visual-wireless data similarity to recognize the social structure within user trajectories. Comprehensive experimental results in indoor and outdoor environments show that the proposed ViWise is robust in trajectory relationship recognition with an accuracy of above 90%.","PeriodicalId":29764,"journal":{"name":"ACM Transactions on Internet of Things","volume":"76 4 1","pages":""},"PeriodicalIF":3.5000,"publicationDate":"2023-08-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"ViWise: Fusing Visual and Wireless Sensing Data for Trajectory Relationship Recognition\",\"authors\":\"Fang-Jing Wu, Sheng-Wun Lai, Sok-Ian Sou\",\"doi\":\"10.1145/3614441\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"People usually form a social structure (e.g., a leader-follower, companion, or independent group) for better interactions among them and thus share similar perceptions of visible scenes and invisible wireless signals encountered while moving. Many mobility-driven applications have paid much attention to recognizing trajectory relationships among people. This work models visual and wireless data to quantify the trajectory similarity between a pair of users. We design a visual and wireless sensor fusion system, called ViWise, which incorporates the first-person video frames collected by a wearable visual device and the wireless packets broadcast by a personal mobile device for recognizing finer-grained trajectory relationships within a mobility group. When people take similar trajectories, they usually share similar visual scenes. Their wireless packets observed by ambient wireless base stations (called wireless scanners in this work) usually contain similar patterns. We model the visual characteristics of physical objects seen by a user from two perspectives: micro-scale image structure with pixel-wise features and macro-scale semantic context. On the other hand, we model characteristics of wireless packets based on the encountered wireless scanners along the user’s trajectory. Given two users’ trajectories, their trajectory characteristics behind the visible video frames and invisible wireless packets are fused together to compute the visual-wireless data similarity that quantifies the correlation between trajectories taken by them. We exploit modeled visual-wireless data similarity to recognize the social structure within user trajectories. Comprehensive experimental results in indoor and outdoor environments show that the proposed ViWise is robust in trajectory relationship recognition with an accuracy of above 90%.\",\"PeriodicalId\":29764,\"journal\":{\"name\":\"ACM Transactions on Internet of Things\",\"volume\":\"76 4 1\",\"pages\":\"\"},\"PeriodicalIF\":3.5000,\"publicationDate\":\"2023-08-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ACM Transactions on Internet of Things\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3614441\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACM Transactions on Internet of Things","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3614441","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
ViWise: Fusing Visual and Wireless Sensing Data for Trajectory Relationship Recognition
People usually form a social structure (e.g., a leader-follower, companion, or independent group) for better interactions among them and thus share similar perceptions of visible scenes and invisible wireless signals encountered while moving. Many mobility-driven applications have paid much attention to recognizing trajectory relationships among people. This work models visual and wireless data to quantify the trajectory similarity between a pair of users. We design a visual and wireless sensor fusion system, called ViWise, which incorporates the first-person video frames collected by a wearable visual device and the wireless packets broadcast by a personal mobile device for recognizing finer-grained trajectory relationships within a mobility group. When people take similar trajectories, they usually share similar visual scenes. Their wireless packets observed by ambient wireless base stations (called wireless scanners in this work) usually contain similar patterns. We model the visual characteristics of physical objects seen by a user from two perspectives: micro-scale image structure with pixel-wise features and macro-scale semantic context. On the other hand, we model characteristics of wireless packets based on the encountered wireless scanners along the user’s trajectory. Given two users’ trajectories, their trajectory characteristics behind the visible video frames and invisible wireless packets are fused together to compute the visual-wireless data similarity that quantifies the correlation between trajectories taken by them. We exploit modeled visual-wireless data similarity to recognize the social structure within user trajectories. Comprehensive experimental results in indoor and outdoor environments show that the proposed ViWise is robust in trajectory relationship recognition with an accuracy of above 90%.