{"title":"Rainbow-RND:一种增强了内在好奇心的基于值的算法","authors":"Sarah Nait Bahloul, Younes Mahmoudi","doi":"10.1109/ICISAT54145.2021.9678409","DOIUrl":null,"url":null,"abstract":"Deep Reinforcement Learning (DRL) is, without a doubt, one of the most promising and exciting research area in Artificial Intelligence (AI). Several approaches have been proposed and improved in a short time to solve different problems. To tackle exploration’s issues, we present in our work a new approach based on curiosity to generate intrinsic rewards. These latter are related to complex environments with sparse rewards, such as the notorious Montezuma’s Revenge and an even more recent and complicated environment named Obstacle Tower. This type of environments requires the agent to generalize its knowledge, learn high-level planning and low-level control. The results of our experimentations showed that a value-based algorithm (such as Rainbow) can be successfully used with a curiosity-based exploration approach (such as random network distillation). This combination effectively performs better in the Obstacle Tower environment.","PeriodicalId":112478,"journal":{"name":"2021 International Conference on Information Systems and Advanced Technologies (ICISAT)","volume":"142 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Rainbow-RND: a Value-based Algorithm Augmented with Intrinsic Curiosity\",\"authors\":\"Sarah Nait Bahloul, Younes Mahmoudi\",\"doi\":\"10.1109/ICISAT54145.2021.9678409\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Deep Reinforcement Learning (DRL) is, without a doubt, one of the most promising and exciting research area in Artificial Intelligence (AI). Several approaches have been proposed and improved in a short time to solve different problems. To tackle exploration’s issues, we present in our work a new approach based on curiosity to generate intrinsic rewards. These latter are related to complex environments with sparse rewards, such as the notorious Montezuma’s Revenge and an even more recent and complicated environment named Obstacle Tower. This type of environments requires the agent to generalize its knowledge, learn high-level planning and low-level control. The results of our experimentations showed that a value-based algorithm (such as Rainbow) can be successfully used with a curiosity-based exploration approach (such as random network distillation). This combination effectively performs better in the Obstacle Tower environment.\",\"PeriodicalId\":112478,\"journal\":{\"name\":\"2021 International Conference on Information Systems and Advanced Technologies (ICISAT)\",\"volume\":\"142 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 International Conference on Information Systems and Advanced Technologies (ICISAT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICISAT54145.2021.9678409\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Information Systems and Advanced Technologies (ICISAT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICISAT54145.2021.9678409","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Rainbow-RND: a Value-based Algorithm Augmented with Intrinsic Curiosity
Deep Reinforcement Learning (DRL) is, without a doubt, one of the most promising and exciting research area in Artificial Intelligence (AI). Several approaches have been proposed and improved in a short time to solve different problems. To tackle exploration’s issues, we present in our work a new approach based on curiosity to generate intrinsic rewards. These latter are related to complex environments with sparse rewards, such as the notorious Montezuma’s Revenge and an even more recent and complicated environment named Obstacle Tower. This type of environments requires the agent to generalize its knowledge, learn high-level planning and low-level control. The results of our experimentations showed that a value-based algorithm (such as Rainbow) can be successfully used with a curiosity-based exploration approach (such as random network distillation). This combination effectively performs better in the Obstacle Tower environment.