Q. Gao, T. Lei, Fei Deng, Zhihao Min, W. Yao, Xiaobin Zhang
{"title":"基于深度强化学习的燃料电池无人机能量管理策略","authors":"Q. Gao, T. Lei, Fei Deng, Zhihao Min, W. Yao, Xiaobin Zhang","doi":"10.1109/ICoPESA54515.2022.9754414","DOIUrl":null,"url":null,"abstract":"Electric propulsion UAV powered by hybrid power system consisting of fuel cells and lithium batteries have attracted significant attention for long endurance and zero emission. Different dynamic characteristics for variable power load demanding which can be stochastically affected by the UAV’s flight air dynamic disturbance are difficult to be modeled with energy management system (EMS). In this paper, a Deep Reinforcement Learning (DRL) algorithm, namely twin-delayed Deep Deterministic policy gradient (TD3), is adopted to derivate EMS for hybrid electric UAV which can avoid performance degradation from uncertainty of power system model and curse of dimensionality of traditional algorithm. The simulation results indicate that the TD3-based DRL strategy is able to coordinate multiple electric power sources based on their natural power characteristics, satisfy different flight profiles of UAV. Furthermore, the performances of TD3, Deep Q-Networks (DQN), Deep Deterministic policy gradient (DDPG) and Dynamic Programming (DP) algorithms with different parameters in EMS of hybrid electric UAV were compared and the effectiveness of the algorithm was verified by digital simulation. Comparative results also illustrate that the proposed TD3 method outperforms other two methods in solving multi-objective optimization energy management problem, in terms of hydrogen consumptions, system efficiency and battery’s state of charge (SOC) sustainability.","PeriodicalId":142509,"journal":{"name":"2022 International Conference on Power Energy Systems and Applications (ICoPESA)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-02-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"A Deep Reinforcement Learning Based Energy Management Strategy for Fuel-Cell Electric UAV\",\"authors\":\"Q. Gao, T. Lei, Fei Deng, Zhihao Min, W. Yao, Xiaobin Zhang\",\"doi\":\"10.1109/ICoPESA54515.2022.9754414\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Electric propulsion UAV powered by hybrid power system consisting of fuel cells and lithium batteries have attracted significant attention for long endurance and zero emission. Different dynamic characteristics for variable power load demanding which can be stochastically affected by the UAV’s flight air dynamic disturbance are difficult to be modeled with energy management system (EMS). In this paper, a Deep Reinforcement Learning (DRL) algorithm, namely twin-delayed Deep Deterministic policy gradient (TD3), is adopted to derivate EMS for hybrid electric UAV which can avoid performance degradation from uncertainty of power system model and curse of dimensionality of traditional algorithm. The simulation results indicate that the TD3-based DRL strategy is able to coordinate multiple electric power sources based on their natural power characteristics, satisfy different flight profiles of UAV. Furthermore, the performances of TD3, Deep Q-Networks (DQN), Deep Deterministic policy gradient (DDPG) and Dynamic Programming (DP) algorithms with different parameters in EMS of hybrid electric UAV were compared and the effectiveness of the algorithm was verified by digital simulation. Comparative results also illustrate that the proposed TD3 method outperforms other two methods in solving multi-objective optimization energy management problem, in terms of hydrogen consumptions, system efficiency and battery’s state of charge (SOC) sustainability.\",\"PeriodicalId\":142509,\"journal\":{\"name\":\"2022 International Conference on Power Energy Systems and Applications (ICoPESA)\",\"volume\":\"15 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-02-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 International Conference on Power Energy Systems and Applications (ICoPESA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICoPESA54515.2022.9754414\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Power Energy Systems and Applications (ICoPESA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICoPESA54515.2022.9754414","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Deep Reinforcement Learning Based Energy Management Strategy for Fuel-Cell Electric UAV
Electric propulsion UAV powered by hybrid power system consisting of fuel cells and lithium batteries have attracted significant attention for long endurance and zero emission. Different dynamic characteristics for variable power load demanding which can be stochastically affected by the UAV’s flight air dynamic disturbance are difficult to be modeled with energy management system (EMS). In this paper, a Deep Reinforcement Learning (DRL) algorithm, namely twin-delayed Deep Deterministic policy gradient (TD3), is adopted to derivate EMS for hybrid electric UAV which can avoid performance degradation from uncertainty of power system model and curse of dimensionality of traditional algorithm. The simulation results indicate that the TD3-based DRL strategy is able to coordinate multiple electric power sources based on their natural power characteristics, satisfy different flight profiles of UAV. Furthermore, the performances of TD3, Deep Q-Networks (DQN), Deep Deterministic policy gradient (DDPG) and Dynamic Programming (DP) algorithms with different parameters in EMS of hybrid electric UAV were compared and the effectiveness of the algorithm was verified by digital simulation. Comparative results also illustrate that the proposed TD3 method outperforms other two methods in solving multi-objective optimization energy management problem, in terms of hydrogen consumptions, system efficiency and battery’s state of charge (SOC) sustainability.