Ayesha Ibrahim , Muhammad Zakir Khan , Muhammad Imran , Hadi Larijani , Qammer H. Abbasi , Muhammad Usman
{"title":"RadSpecFusion:多雷达人体活动识别的动态注意力加权","authors":"Ayesha Ibrahim , Muhammad Zakir Khan , Muhammad Imran , Hadi Larijani , Qammer H. Abbasi , Muhammad Usman","doi":"10.1016/j.iot.2025.101682","DOIUrl":null,"url":null,"abstract":"<div><div>This paper presents RadSpecFusion, a novel dynamic attention-based fusion architecture for multi-radar human activity recognition (HAR). Our method learns activity-specific importance weights for each radar modality (24 GHz, 77 GHz, and Xethru sensors). Unlike existing concatenation or averaging approaches, our method dynamically adapts radar contributions based on motion characteristics. This addresses cross-frequency generalization challenges, where transfer learning methods achieve only 11%–34% accuracy. Using the CI4R dataset with spectrograms from 11 activities, our approach achieves 99.21% accuracy, representing a 15.8% improvement over existing fusion methods (83.4%). This demonstrates that different radar frequencies capture complementary information about human motion. Ablation studies show that while the three-radar system optimizes performance, dual-radar combinations achieve comparable accuracy (24GHz+77GHz: 96.1%, 24GHz+Xethru: 95.8%, 77GHz+Xethru: 97.2%), enabling flexible deployment for resource-constrained applications. The attention mechanism reveals interpretable patterns: 77 GHz radar receives higher weights for fine movements (superior Doppler resolution), while 24 GHz dominates gross body movements (better range resolution). The system maintains 71.4% accuracy at 10 dB SNR, demonstrating environmental robustness. This research establishes a new paradigm for multimodal radar fusion, moving from cross-frequency transfer learning to adaptive fusion with implications for healthcare monitoring, smart environments, and security applications.</div></div>","PeriodicalId":29968,"journal":{"name":"Internet of Things","volume":"33 ","pages":"Article 101682"},"PeriodicalIF":7.6000,"publicationDate":"2025-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"RadSpecFusion: Dynamic attention weighting for multi-radar human activity recognition\",\"authors\":\"Ayesha Ibrahim , Muhammad Zakir Khan , Muhammad Imran , Hadi Larijani , Qammer H. Abbasi , Muhammad Usman\",\"doi\":\"10.1016/j.iot.2025.101682\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>This paper presents RadSpecFusion, a novel dynamic attention-based fusion architecture for multi-radar human activity recognition (HAR). Our method learns activity-specific importance weights for each radar modality (24 GHz, 77 GHz, and Xethru sensors). Unlike existing concatenation or averaging approaches, our method dynamically adapts radar contributions based on motion characteristics. This addresses cross-frequency generalization challenges, where transfer learning methods achieve only 11%–34% accuracy. Using the CI4R dataset with spectrograms from 11 activities, our approach achieves 99.21% accuracy, representing a 15.8% improvement over existing fusion methods (83.4%). This demonstrates that different radar frequencies capture complementary information about human motion. Ablation studies show that while the three-radar system optimizes performance, dual-radar combinations achieve comparable accuracy (24GHz+77GHz: 96.1%, 24GHz+Xethru: 95.8%, 77GHz+Xethru: 97.2%), enabling flexible deployment for resource-constrained applications. The attention mechanism reveals interpretable patterns: 77 GHz radar receives higher weights for fine movements (superior Doppler resolution), while 24 GHz dominates gross body movements (better range resolution). The system maintains 71.4% accuracy at 10 dB SNR, demonstrating environmental robustness. This research establishes a new paradigm for multimodal radar fusion, moving from cross-frequency transfer learning to adaptive fusion with implications for healthcare monitoring, smart environments, and security applications.</div></div>\",\"PeriodicalId\":29968,\"journal\":{\"name\":\"Internet of Things\",\"volume\":\"33 \",\"pages\":\"Article 101682\"},\"PeriodicalIF\":7.6000,\"publicationDate\":\"2025-07-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Internet of Things\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2542660525001969\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Internet of Things","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2542660525001969","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
RadSpecFusion: Dynamic attention weighting for multi-radar human activity recognition
This paper presents RadSpecFusion, a novel dynamic attention-based fusion architecture for multi-radar human activity recognition (HAR). Our method learns activity-specific importance weights for each radar modality (24 GHz, 77 GHz, and Xethru sensors). Unlike existing concatenation or averaging approaches, our method dynamically adapts radar contributions based on motion characteristics. This addresses cross-frequency generalization challenges, where transfer learning methods achieve only 11%–34% accuracy. Using the CI4R dataset with spectrograms from 11 activities, our approach achieves 99.21% accuracy, representing a 15.8% improvement over existing fusion methods (83.4%). This demonstrates that different radar frequencies capture complementary information about human motion. Ablation studies show that while the three-radar system optimizes performance, dual-radar combinations achieve comparable accuracy (24GHz+77GHz: 96.1%, 24GHz+Xethru: 95.8%, 77GHz+Xethru: 97.2%), enabling flexible deployment for resource-constrained applications. The attention mechanism reveals interpretable patterns: 77 GHz radar receives higher weights for fine movements (superior Doppler resolution), while 24 GHz dominates gross body movements (better range resolution). The system maintains 71.4% accuracy at 10 dB SNR, demonstrating environmental robustness. This research establishes a new paradigm for multimodal radar fusion, moving from cross-frequency transfer learning to adaptive fusion with implications for healthcare monitoring, smart environments, and security applications.
期刊介绍:
Internet of Things; Engineering Cyber Physical Human Systems is a comprehensive journal encouraging cross collaboration between researchers, engineers and practitioners in the field of IoT & Cyber Physical Human Systems. The journal offers a unique platform to exchange scientific information on the entire breadth of technology, science, and societal applications of the IoT.
The journal will place a high priority on timely publication, and provide a home for high quality.
Furthermore, IOT is interested in publishing topical Special Issues on any aspect of IOT.