{"title":"基于多智能体离散软角色评价算法的多用户协同抗干扰策略","authors":"Xiaorong Jing;Rui Wang;Hongjiang Lei;Hongqing Liu;Qianbin Chen","doi":"10.1109/TIFS.2025.3570160","DOIUrl":null,"url":null,"abstract":"In multi-user adversarial scenarios involving external malicious jamming and internal co-channel interference, environmental instability and increased decision-making dimensions cause traditional deep reinforcement learning (DRL)-based anti-jamming schemes to suffer from insufficient exploration. Agents must choose policies from a large action set, leading to a significant decline in anti-jamming performance. To address these issues, this paper proposes a multi-agent discrete soft actor-critic (MA-DSAC) algorithm-based collaborative anti-jamming strategy, integrating frequency, power, and modulation-coding domains. This strategy first introduces a Markov game to model and analyze the multi-user anti-jamming problem. Next, the soft actor-critic (SAC) algorithm is discretized to handle the multi-dimensional discrete action space. Finally, through information exchange between communication transceivers and based on a centralized training with decentralized execution (CTDE) framework, it is extended to a multi-agent DRL algorithm to achieve efficient multi-user cooperative anti-jamming. Simulation results show that in various anti-jamming scenarios with both fixed-mode and intelligent jammers, the proposed anti-jamming strategy’s performance improves by more than 25% compared to traditional value-based DRL strategies, including independent deep Q-network (I-DQN) and multi-agent virtual exploration in deep Q-learning (MA-VEDQL). Furthermore, through information exchange between communication transceivers, the instability problem of multi-agent DRL is effectively alleviated, enabling the communication transceivers to balance competition and cooperation. Consequently, its anti-jamming performance improves by more than 6% compared to the independent DSAC (I-DSAC) strategy.","PeriodicalId":13492,"journal":{"name":"IEEE Transactions on Information Forensics and Security","volume":"20 ","pages":"5025-5038"},"PeriodicalIF":8.0000,"publicationDate":"2025-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multi-Agent Discrete Soft Actor-Critic Algorithm-Based Multi-User Collaborative Anti-Jamming Strategy\",\"authors\":\"Xiaorong Jing;Rui Wang;Hongjiang Lei;Hongqing Liu;Qianbin Chen\",\"doi\":\"10.1109/TIFS.2025.3570160\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In multi-user adversarial scenarios involving external malicious jamming and internal co-channel interference, environmental instability and increased decision-making dimensions cause traditional deep reinforcement learning (DRL)-based anti-jamming schemes to suffer from insufficient exploration. Agents must choose policies from a large action set, leading to a significant decline in anti-jamming performance. To address these issues, this paper proposes a multi-agent discrete soft actor-critic (MA-DSAC) algorithm-based collaborative anti-jamming strategy, integrating frequency, power, and modulation-coding domains. This strategy first introduces a Markov game to model and analyze the multi-user anti-jamming problem. Next, the soft actor-critic (SAC) algorithm is discretized to handle the multi-dimensional discrete action space. Finally, through information exchange between communication transceivers and based on a centralized training with decentralized execution (CTDE) framework, it is extended to a multi-agent DRL algorithm to achieve efficient multi-user cooperative anti-jamming. Simulation results show that in various anti-jamming scenarios with both fixed-mode and intelligent jammers, the proposed anti-jamming strategy’s performance improves by more than 25% compared to traditional value-based DRL strategies, including independent deep Q-network (I-DQN) and multi-agent virtual exploration in deep Q-learning (MA-VEDQL). Furthermore, through information exchange between communication transceivers, the instability problem of multi-agent DRL is effectively alleviated, enabling the communication transceivers to balance competition and cooperation. Consequently, its anti-jamming performance improves by more than 6% compared to the independent DSAC (I-DSAC) strategy.\",\"PeriodicalId\":13492,\"journal\":{\"name\":\"IEEE Transactions on Information Forensics and Security\",\"volume\":\"20 \",\"pages\":\"5025-5038\"},\"PeriodicalIF\":8.0000,\"publicationDate\":\"2025-03-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Information Forensics and Security\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/11003932/\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, THEORY & METHODS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Information Forensics and Security","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/11003932/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
In multi-user adversarial scenarios involving external malicious jamming and internal co-channel interference, environmental instability and increased decision-making dimensions cause traditional deep reinforcement learning (DRL)-based anti-jamming schemes to suffer from insufficient exploration. Agents must choose policies from a large action set, leading to a significant decline in anti-jamming performance. To address these issues, this paper proposes a multi-agent discrete soft actor-critic (MA-DSAC) algorithm-based collaborative anti-jamming strategy, integrating frequency, power, and modulation-coding domains. This strategy first introduces a Markov game to model and analyze the multi-user anti-jamming problem. Next, the soft actor-critic (SAC) algorithm is discretized to handle the multi-dimensional discrete action space. Finally, through information exchange between communication transceivers and based on a centralized training with decentralized execution (CTDE) framework, it is extended to a multi-agent DRL algorithm to achieve efficient multi-user cooperative anti-jamming. Simulation results show that in various anti-jamming scenarios with both fixed-mode and intelligent jammers, the proposed anti-jamming strategy’s performance improves by more than 25% compared to traditional value-based DRL strategies, including independent deep Q-network (I-DQN) and multi-agent virtual exploration in deep Q-learning (MA-VEDQL). Furthermore, through information exchange between communication transceivers, the instability problem of multi-agent DRL is effectively alleviated, enabling the communication transceivers to balance competition and cooperation. Consequently, its anti-jamming performance improves by more than 6% compared to the independent DSAC (I-DSAC) strategy.
期刊介绍:
The IEEE Transactions on Information Forensics and Security covers the sciences, technologies, and applications relating to information forensics, information security, biometrics, surveillance and systems applications that incorporate these features