STAR-RIS辅助保密通信与深度强化学习

IF 5.3 2区 计算机科学 Q1 TELECOMMUNICATIONS
Miao Zhang;Xuran Ding;Yanqun Tang;Shixun Wu;Kai Xu
{"title":"STAR-RIS辅助保密通信与深度强化学习","authors":"Miao Zhang;Xuran Ding;Yanqun Tang;Shixun Wu;Kai Xu","doi":"10.1109/TGCN.2024.3466189","DOIUrl":null,"url":null,"abstract":"In this paper, we investigate the secure transmission in a simultaneously transmitting and reflecting reconfigurable intelligent surface (STAR-RIS)-assisted down link multiple-input single-output (MISO) wireless network. The secrecy rate is maximized by the joint design of the transmit beamforming, the transmission and reflection coefficients of the STAR-RIS, while satisfying the electromagnetic property of the STAR-RIS and transmit power limit of the base station. Since this communication network is in a dynamic environment, the optimization problem is non-convex and mathematically difficult to solve. To address this issue, two deep reinforcement learning (DRL)-based algorithms, namely soft actor-critic (SAC) algorithm and soft actor-critic based on loss-adjusted approximate actor prioritized experience replay (L3APER-SAC) are proposed to obtain the maximum reward by constantly interacting with and learning from the dynamic environment. Moreover, for the L3APER-SAC algorithm, to achieve higher performance and stability, we introduce two experience replay buffers—one is regular experience replay and the other is prioritized experience replay. Simulation results comprehensively assess the performance of two DRL algorithms and indicate that both proposed algorithms outperform benchmark approaches. Particularly, L3APER-SAC, exhibits superior performance, albeit with an associated increase in computational complexity.","PeriodicalId":13052,"journal":{"name":"IEEE Transactions on Green Communications and Networking","volume":"9 2","pages":"739-753"},"PeriodicalIF":5.3000,"publicationDate":"2024-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"STAR-RIS Assisted Secrecy Communication With Deep Reinforcement Learning\",\"authors\":\"Miao Zhang;Xuran Ding;Yanqun Tang;Shixun Wu;Kai Xu\",\"doi\":\"10.1109/TGCN.2024.3466189\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, we investigate the secure transmission in a simultaneously transmitting and reflecting reconfigurable intelligent surface (STAR-RIS)-assisted down link multiple-input single-output (MISO) wireless network. The secrecy rate is maximized by the joint design of the transmit beamforming, the transmission and reflection coefficients of the STAR-RIS, while satisfying the electromagnetic property of the STAR-RIS and transmit power limit of the base station. Since this communication network is in a dynamic environment, the optimization problem is non-convex and mathematically difficult to solve. To address this issue, two deep reinforcement learning (DRL)-based algorithms, namely soft actor-critic (SAC) algorithm and soft actor-critic based on loss-adjusted approximate actor prioritized experience replay (L3APER-SAC) are proposed to obtain the maximum reward by constantly interacting with and learning from the dynamic environment. Moreover, for the L3APER-SAC algorithm, to achieve higher performance and stability, we introduce two experience replay buffers—one is regular experience replay and the other is prioritized experience replay. Simulation results comprehensively assess the performance of two DRL algorithms and indicate that both proposed algorithms outperform benchmark approaches. Particularly, L3APER-SAC, exhibits superior performance, albeit with an associated increase in computational complexity.\",\"PeriodicalId\":13052,\"journal\":{\"name\":\"IEEE Transactions on Green Communications and Networking\",\"volume\":\"9 2\",\"pages\":\"739-753\"},\"PeriodicalIF\":5.3000,\"publicationDate\":\"2024-09-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Green Communications and Networking\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10689376/\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"TELECOMMUNICATIONS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Green Communications and Networking","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10689376/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"TELECOMMUNICATIONS","Score":null,"Total":0}
引用次数: 0

摘要

本文研究了同时发射和反射可重构智能面(STAR-RIS)辅助下链路多输入单输出(MISO)无线网络中的安全传输问题。在满足星- ris的电磁特性和基站发射功率限制的前提下,通过星- ris发射波束形成、发射和反射系数的联合设计,使保密率达到最大。由于该通信网络处于动态环境中,优化问题是非凸的,在数学上难以求解。为了解决这一问题,提出了两种基于深度强化学习(DRL)的算法,即软行为者评价(SAC)算法和基于损失调整的近似行为者优先经验重播(L3APER-SAC)算法,通过不断与动态环境交互和学习来获得最大的奖励。此外,对于L3APER-SAC算法,为了获得更高的性能和稳定性,我们引入了两种经验重播缓冲器,一种是常规经验重播缓冲器,另一种是优先经验重播缓冲器。仿真结果全面评估了两种DRL算法的性能,并表明两种算法都优于基准方法。特别是,L3APER-SAC表现出了卓越的性能,尽管计算复杂性有所增加。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
STAR-RIS Assisted Secrecy Communication With Deep Reinforcement Learning
In this paper, we investigate the secure transmission in a simultaneously transmitting and reflecting reconfigurable intelligent surface (STAR-RIS)-assisted down link multiple-input single-output (MISO) wireless network. The secrecy rate is maximized by the joint design of the transmit beamforming, the transmission and reflection coefficients of the STAR-RIS, while satisfying the electromagnetic property of the STAR-RIS and transmit power limit of the base station. Since this communication network is in a dynamic environment, the optimization problem is non-convex and mathematically difficult to solve. To address this issue, two deep reinforcement learning (DRL)-based algorithms, namely soft actor-critic (SAC) algorithm and soft actor-critic based on loss-adjusted approximate actor prioritized experience replay (L3APER-SAC) are proposed to obtain the maximum reward by constantly interacting with and learning from the dynamic environment. Moreover, for the L3APER-SAC algorithm, to achieve higher performance and stability, we introduce two experience replay buffers—one is regular experience replay and the other is prioritized experience replay. Simulation results comprehensively assess the performance of two DRL algorithms and indicate that both proposed algorithms outperform benchmark approaches. Particularly, L3APER-SAC, exhibits superior performance, albeit with an associated increase in computational complexity.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
IEEE Transactions on Green Communications and Networking
IEEE Transactions on Green Communications and Networking Computer Science-Computer Networks and Communications
CiteScore
9.30
自引率
6.20%
发文量
181
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信