{"title":"基于变长时间步长多智能体强化学习的传感器网络协同充电框架","authors":"Yuxin Chen, He Wu, Yongheng Liang, G. Lai","doi":"10.1109/SECON52354.2021.9491594","DOIUrl":null,"url":null,"abstract":"This paper studies cooperative charging, in which multiple mobile chargers cooperatively provide wireless charging services in a Wireless Rechargeable Sensor Network (WRSN). The ultimate goal of this cooperative charging is the long-term optimization that maximizes both the lifetime of all sensor nodes and the charging utility of each Mobile Charger (MC). We have attempted to apply Multi-Agent Reinforcement Learning (MARL) algorithms to this problem. Unfortunately, similar to existing methods, MARL algorithms also fail early in cooperative charging. We found that an MARL algorithm trained in each time-step of fixed length is neither accurate nor efficient in cooperative charging. We propose a new MARL framework, called VarLenMARL. For the accuracy of reward estimation, VarLenMARL allows each MC completes an action within a time-step of variable length before estimating rewards. Furthermore, we design a special mechanism in VarLenMARL for the long-term optimality of cooperative charging within a WRSN. Our results show that algorithms implemented on VarLenMARL achieved both higher charging utility of MCs and longer lifetime of sensor nodes.","PeriodicalId":120945,"journal":{"name":"2021 18th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"VarLenMARL: A Framework of Variable-Length Time-Step Multi-Agent Reinforcement Learning for Cooperative Charging in Sensor Networks\",\"authors\":\"Yuxin Chen, He Wu, Yongheng Liang, G. Lai\",\"doi\":\"10.1109/SECON52354.2021.9491594\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper studies cooperative charging, in which multiple mobile chargers cooperatively provide wireless charging services in a Wireless Rechargeable Sensor Network (WRSN). The ultimate goal of this cooperative charging is the long-term optimization that maximizes both the lifetime of all sensor nodes and the charging utility of each Mobile Charger (MC). We have attempted to apply Multi-Agent Reinforcement Learning (MARL) algorithms to this problem. Unfortunately, similar to existing methods, MARL algorithms also fail early in cooperative charging. We found that an MARL algorithm trained in each time-step of fixed length is neither accurate nor efficient in cooperative charging. We propose a new MARL framework, called VarLenMARL. For the accuracy of reward estimation, VarLenMARL allows each MC completes an action within a time-step of variable length before estimating rewards. Furthermore, we design a special mechanism in VarLenMARL for the long-term optimality of cooperative charging within a WRSN. Our results show that algorithms implemented on VarLenMARL achieved both higher charging utility of MCs and longer lifetime of sensor nodes.\",\"PeriodicalId\":120945,\"journal\":{\"name\":\"2021 18th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON)\",\"volume\":\"18 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-07-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 18th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SECON52354.2021.9491594\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 18th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SECON52354.2021.9491594","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
VarLenMARL: A Framework of Variable-Length Time-Step Multi-Agent Reinforcement Learning for Cooperative Charging in Sensor Networks
This paper studies cooperative charging, in which multiple mobile chargers cooperatively provide wireless charging services in a Wireless Rechargeable Sensor Network (WRSN). The ultimate goal of this cooperative charging is the long-term optimization that maximizes both the lifetime of all sensor nodes and the charging utility of each Mobile Charger (MC). We have attempted to apply Multi-Agent Reinforcement Learning (MARL) algorithms to this problem. Unfortunately, similar to existing methods, MARL algorithms also fail early in cooperative charging. We found that an MARL algorithm trained in each time-step of fixed length is neither accurate nor efficient in cooperative charging. We propose a new MARL framework, called VarLenMARL. For the accuracy of reward estimation, VarLenMARL allows each MC completes an action within a time-step of variable length before estimating rewards. Furthermore, we design a special mechanism in VarLenMARL for the long-term optimality of cooperative charging within a WRSN. Our results show that algorithms implemented on VarLenMARL achieved both higher charging utility of MCs and longer lifetime of sensor nodes.