一种基于强化学习的多导弹协同制导律

Hongxu Chen, Jianglong Yu, Xiwang Dong
{"title":"一种基于强化学习的多导弹协同制导律","authors":"Hongxu Chen, Jianglong Yu, Xiwang Dong","doi":"10.1109/ICUS55513.2022.9986718","DOIUrl":null,"url":null,"abstract":"The traditional proportional guidance law lacks the limitation of time and field of view. In order to realize the coordinated attack of multiple missiles on targets and improve the attack efficiency, a reinforcement learning cooperative guidance law based on deep deterministic policy gradient descent neural network is proposed. According to the particularity of guidance process, the reinforcement learning agent is obtained by constructing state space, action space and reward function training. The simulation results show that the enhanced learning guidance law can strike maneuvering targets simultaneously and satisfy the field of view constraint, which is superior to the traditional cooperative proportional guidance law.","PeriodicalId":345773,"journal":{"name":"2022 IEEE International Conference on Unmanned Systems (ICUS)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"A Cooperative Guidance Law for Multiple Missiles based on Reinforcement Learning\",\"authors\":\"Hongxu Chen, Jianglong Yu, Xiwang Dong\",\"doi\":\"10.1109/ICUS55513.2022.9986718\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The traditional proportional guidance law lacks the limitation of time and field of view. In order to realize the coordinated attack of multiple missiles on targets and improve the attack efficiency, a reinforcement learning cooperative guidance law based on deep deterministic policy gradient descent neural network is proposed. According to the particularity of guidance process, the reinforcement learning agent is obtained by constructing state space, action space and reward function training. The simulation results show that the enhanced learning guidance law can strike maneuvering targets simultaneously and satisfy the field of view constraint, which is superior to the traditional cooperative proportional guidance law.\",\"PeriodicalId\":345773,\"journal\":{\"name\":\"2022 IEEE International Conference on Unmanned Systems (ICUS)\",\"volume\":\"2 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-10-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE International Conference on Unmanned Systems (ICUS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICUS55513.2022.9986718\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE International Conference on Unmanned Systems (ICUS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICUS55513.2022.9986718","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

传统的比例导引法缺乏时间和视场的限制。为了实现多枚导弹对目标的协同攻击,提高攻击效率,提出了一种基于深度确定性策略梯度下降神经网络的强化学习协同制导律。根据引导过程的特殊性,通过构造状态空间、动作空间和奖励函数训练获得强化学习智能体。仿真结果表明,改进的学习制导律能够同时打击机动目标并满足视场约束,优于传统的协同比例制导律。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A Cooperative Guidance Law for Multiple Missiles based on Reinforcement Learning
The traditional proportional guidance law lacks the limitation of time and field of view. In order to realize the coordinated attack of multiple missiles on targets and improve the attack efficiency, a reinforcement learning cooperative guidance law based on deep deterministic policy gradient descent neural network is proposed. According to the particularity of guidance process, the reinforcement learning agent is obtained by constructing state space, action space and reward function training. The simulation results show that the enhanced learning guidance law can strike maneuvering targets simultaneously and satisfy the field of view constraint, which is superior to the traditional cooperative proportional guidance law.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信