{"title":"无线网络中自主功率配置与控制的间接强化学习","authors":"A. Udenze, K. Mcdonald-Maier","doi":"10.1109/AHS.2009.51","DOIUrl":null,"url":null,"abstract":"In this paper, non deterministic Indirect Reinforcement Learning (RL) techniques for controlling the transmission times and power of Wireless Network nodes are presented. Indirect RL facilitates planning and learning which ultimately leads to convergence on optimal actions with reduced episodes or time steps compared to direct RL. Three Dyna architecture based algorithms for non deterministic environments are presented. The results show improvements over direct RL and conventional static power control techniques.","PeriodicalId":318989,"journal":{"name":"2009 NASA/ESA Conference on Adaptive Hardware and Systems","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Indirect Reinforcement Learning for Autonomous Power Configuration and Control in Wireless Networks\",\"authors\":\"A. Udenze, K. Mcdonald-Maier\",\"doi\":\"10.1109/AHS.2009.51\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, non deterministic Indirect Reinforcement Learning (RL) techniques for controlling the transmission times and power of Wireless Network nodes are presented. Indirect RL facilitates planning and learning which ultimately leads to convergence on optimal actions with reduced episodes or time steps compared to direct RL. Three Dyna architecture based algorithms for non deterministic environments are presented. The results show improvements over direct RL and conventional static power control techniques.\",\"PeriodicalId\":318989,\"journal\":{\"name\":\"2009 NASA/ESA Conference on Adaptive Hardware and Systems\",\"volume\":\"8 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1900-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2009 NASA/ESA Conference on Adaptive Hardware and Systems\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/AHS.2009.51\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2009 NASA/ESA Conference on Adaptive Hardware and Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/AHS.2009.51","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Indirect Reinforcement Learning for Autonomous Power Configuration and Control in Wireless Networks
In this paper, non deterministic Indirect Reinforcement Learning (RL) techniques for controlling the transmission times and power of Wireless Network nodes are presented. Indirect RL facilitates planning and learning which ultimately leads to convergence on optimal actions with reduced episodes or time steps compared to direct RL. Three Dyna architecture based algorithms for non deterministic environments are presented. The results show improvements over direct RL and conventional static power control techniques.