Q-learning vs. FRIQ-learning in the Maze problem

T. Tompa, S. Kovács
{"title":"Q-learning vs. FRIQ-learning in the Maze problem","authors":"T. Tompa, S. Kovács","doi":"10.1109/COGINFOCOM.2015.7390652","DOIUrl":null,"url":null,"abstract":"The goal of this paper is to give a demonstrative example for introducing the benefits of the FRIQ-learning (Fuzzy Rule Interpolation-based Q-learning) versus the traditional discrete Q-learning. The chosen example is an easily scalable discrete state and discrete action space task the Maze problem. The main difference of the two studied reinforcement learning methods, that the traditional Q-learning has discrete state, action and Q-function representation. While the FRIQ-learning has continuous state, action space and a Fuzzy Rule Interpolation based Q-function representation. For comparing the convergence speed of the two methods, both will start from an empty knowledge base, zero Q-table for the Q-learning and empty rule-base for the FRIQ-learning and following the same policy stops at the same performance condition. In the example of the paper the Maze problem will be studied in different obstacle configurations and different scaling.","PeriodicalId":377891,"journal":{"name":"2015 6th IEEE International Conference on Cognitive Infocommunications (CogInfoCom)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2015-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2015 6th IEEE International Conference on Cognitive Infocommunications (CogInfoCom)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/COGINFOCOM.2015.7390652","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

The goal of this paper is to give a demonstrative example for introducing the benefits of the FRIQ-learning (Fuzzy Rule Interpolation-based Q-learning) versus the traditional discrete Q-learning. The chosen example is an easily scalable discrete state and discrete action space task the Maze problem. The main difference of the two studied reinforcement learning methods, that the traditional Q-learning has discrete state, action and Q-function representation. While the FRIQ-learning has continuous state, action space and a Fuzzy Rule Interpolation based Q-function representation. For comparing the convergence speed of the two methods, both will start from an empty knowledge base, zero Q-table for the Q-learning and empty rule-base for the FRIQ-learning and following the same policy stops at the same performance condition. In the example of the paper the Maze problem will be studied in different obstacle configurations and different scaling.
迷宫问题中的q -学习和friq -学习
本文的目的是给出一个示范性的例子,介绍friq学习(基于模糊规则插值的q学习)与传统离散q学习的优势。所选择的例子是一个易于伸缩的离散状态和离散动作空间任务的迷宫问题。两种强化学习方法的主要区别在于传统的q -学习具有离散的状态、动作和q -函数表示。而friq学习具有连续状态、动作空间和基于模糊规则插值的q函数表示。为了比较两种方法的收敛速度,两种方法都从一个空的知识库开始,Q-learning的q表为零,FRIQ-learning的规则库为空,遵循相同的策略,在相同的性能条件下停止。在本文的例子中,我们将研究不同障碍物配置和不同尺度下的迷宫问题。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信