Che Wang, Jifeng Hu, Fuhu Song, Jiao Huang, Zixuan Yang, Yusen Wang
{"title":"Multiple Frequency Bands Temporal State Representation for Deep Reinforcement Learning","authors":"Che Wang, Jifeng Hu, Fuhu Song, Jiao Huang, Zixuan Yang, Yusen Wang","doi":"10.1145/3590003.3590058","DOIUrl":null,"url":null,"abstract":"Deep reinforcement learning has achieved significant success in solving sequential decision-making tasks. Excellent models usually require the input of valid state signals during training, which is challenging to encode temporal state features for the deep reinforcement learning model. To address this issue, recent methods attempt to encode multi-step sequential state signals so as to obtain more comprehensive observational information. However, these methods usually have a lower performance on complex continuous control tasks because mapping the state sequence into a low-dimensional embedding causes blurring of the immediate state features. In this paper, we propose a multiple frequency bands temporal state representation learning framework. The temporal state signals are decomposed into discrete state signals of various frequency bands by Discrete Fourier Transform (DFT). Then, feature signals filtered out different high-frequency bands are generated. Meanwhile, the mask generator evaluates the weights of signals of various frequency bands and encodes high-quality representations for agent training. Our intuition is that temporal state representations considering multiple frequency bands have high fidelity and stability. We conduct experiments tasks and verify that our method has obvious advantages over the baseline in complex continuous control tasks such as Walker and Crawler.","PeriodicalId":340225,"journal":{"name":"Proceedings of the 2023 2nd Asia Conference on Algorithms, Computing and Machine Learning","volume":"6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2023 2nd Asia Conference on Algorithms, Computing and Machine Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3590003.3590058","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Deep reinforcement learning has achieved significant success in solving sequential decision-making tasks. Excellent models usually require the input of valid state signals during training, which is challenging to encode temporal state features for the deep reinforcement learning model. To address this issue, recent methods attempt to encode multi-step sequential state signals so as to obtain more comprehensive observational information. However, these methods usually have a lower performance on complex continuous control tasks because mapping the state sequence into a low-dimensional embedding causes blurring of the immediate state features. In this paper, we propose a multiple frequency bands temporal state representation learning framework. The temporal state signals are decomposed into discrete state signals of various frequency bands by Discrete Fourier Transform (DFT). Then, feature signals filtered out different high-frequency bands are generated. Meanwhile, the mask generator evaluates the weights of signals of various frequency bands and encodes high-quality representations for agent training. Our intuition is that temporal state representations considering multiple frequency bands have high fidelity and stability. We conduct experiments tasks and verify that our method has obvious advantages over the baseline in complex continuous control tasks such as Walker and Crawler.