{"title":"Q-learning vs. FRIQ-learning in the Maze problem","authors":"T. Tompa, S. Kovács","doi":"10.1109/COGINFOCOM.2015.7390652","DOIUrl":null,"url":null,"abstract":"The goal of this paper is to give a demonstrative example for introducing the benefits of the FRIQ-learning (Fuzzy Rule Interpolation-based Q-learning) versus the traditional discrete Q-learning. The chosen example is an easily scalable discrete state and discrete action space task the Maze problem. The main difference of the two studied reinforcement learning methods, that the traditional Q-learning has discrete state, action and Q-function representation. While the FRIQ-learning has continuous state, action space and a Fuzzy Rule Interpolation based Q-function representation. For comparing the convergence speed of the two methods, both will start from an empty knowledge base, zero Q-table for the Q-learning and empty rule-base for the FRIQ-learning and following the same policy stops at the same performance condition. In the example of the paper the Maze problem will be studied in different obstacle configurations and different scaling.","PeriodicalId":377891,"journal":{"name":"2015 6th IEEE International Conference on Cognitive Infocommunications (CogInfoCom)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2015-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2015 6th IEEE International Conference on Cognitive Infocommunications (CogInfoCom)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/COGINFOCOM.2015.7390652","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
The goal of this paper is to give a demonstrative example for introducing the benefits of the FRIQ-learning (Fuzzy Rule Interpolation-based Q-learning) versus the traditional discrete Q-learning. The chosen example is an easily scalable discrete state and discrete action space task the Maze problem. The main difference of the two studied reinforcement learning methods, that the traditional Q-learning has discrete state, action and Q-function representation. While the FRIQ-learning has continuous state, action space and a Fuzzy Rule Interpolation based Q-function representation. For comparing the convergence speed of the two methods, both will start from an empty knowledge base, zero Q-table for the Q-learning and empty rule-base for the FRIQ-learning and following the same policy stops at the same performance condition. In the example of the paper the Maze problem will be studied in different obstacle configurations and different scaling.