{"title":"竞争性Takagi-Sugeno模糊强化学习","authors":"X.W. Yan, Z. Deng, Z. Sun","doi":"10.1109/CCA.2001.973980","DOIUrl":null,"url":null,"abstract":"This paper proposes a competitive Takagi-Sugeno fuzzy reinforcement learning network (CTSFRLN) for solving complicated learning tasks of continuous domains. The proposed CTSFRLN is constructed by combining Takagi-Sugeno type fuzzy inference systems with action-value-based reinforcement learning methods. The architecture of CTSFRLN is described and a fitting exploration strategy, i.e., max-min Boltzmann exploration, is developed to implement local competitions in rule consequents. Three competitive learning algorithms are derived, including the competitive Takagi-Sugeno fuzzy Q-learning, competitive Takagi-Sugeno fuzzy R-learning, and competitive Takagi-Sugeno fuzzy advantage learning. These learning methods lead to the so called Takagi-Sugeno fuzzy variable structure controller. Experiments on the double inverted pendulum system demonstrate the performance and applicability of the proposed schemes. The superiority of these methods with respect to other related reinforcement learning ones is also illustrated. Finally, the conclusion remark is drawn.","PeriodicalId":365390,"journal":{"name":"Proceedings of the 2001 IEEE International Conference on Control Applications (CCA'01) (Cat. No.01CH37204)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2001-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":"{\"title\":\"Competitive Takagi-Sugeno fuzzy reinforcement learning\",\"authors\":\"X.W. Yan, Z. Deng, Z. Sun\",\"doi\":\"10.1109/CCA.2001.973980\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper proposes a competitive Takagi-Sugeno fuzzy reinforcement learning network (CTSFRLN) for solving complicated learning tasks of continuous domains. The proposed CTSFRLN is constructed by combining Takagi-Sugeno type fuzzy inference systems with action-value-based reinforcement learning methods. The architecture of CTSFRLN is described and a fitting exploration strategy, i.e., max-min Boltzmann exploration, is developed to implement local competitions in rule consequents. Three competitive learning algorithms are derived, including the competitive Takagi-Sugeno fuzzy Q-learning, competitive Takagi-Sugeno fuzzy R-learning, and competitive Takagi-Sugeno fuzzy advantage learning. These learning methods lead to the so called Takagi-Sugeno fuzzy variable structure controller. Experiments on the double inverted pendulum system demonstrate the performance and applicability of the proposed schemes. The superiority of these methods with respect to other related reinforcement learning ones is also illustrated. Finally, the conclusion remark is drawn.\",\"PeriodicalId\":365390,\"journal\":{\"name\":\"Proceedings of the 2001 IEEE International Conference on Control Applications (CCA'01) (Cat. No.01CH37204)\",\"volume\":\"4 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2001-09-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"6\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2001 IEEE International Conference on Control Applications (CCA'01) (Cat. No.01CH37204)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CCA.2001.973980\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2001 IEEE International Conference on Control Applications (CCA'01) (Cat. No.01CH37204)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CCA.2001.973980","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
This paper proposes a competitive Takagi-Sugeno fuzzy reinforcement learning network (CTSFRLN) for solving complicated learning tasks of continuous domains. The proposed CTSFRLN is constructed by combining Takagi-Sugeno type fuzzy inference systems with action-value-based reinforcement learning methods. The architecture of CTSFRLN is described and a fitting exploration strategy, i.e., max-min Boltzmann exploration, is developed to implement local competitions in rule consequents. Three competitive learning algorithms are derived, including the competitive Takagi-Sugeno fuzzy Q-learning, competitive Takagi-Sugeno fuzzy R-learning, and competitive Takagi-Sugeno fuzzy advantage learning. These learning methods lead to the so called Takagi-Sugeno fuzzy variable structure controller. Experiments on the double inverted pendulum system demonstrate the performance and applicability of the proposed schemes. The superiority of these methods with respect to other related reinforcement learning ones is also illustrated. Finally, the conclusion remark is drawn.