mp -草稿:排序搜索树和精炼游戏棋盘表示以改进草稿的多代理系统

V. Duarte, Rita Maria Silva Julia
{"title":"mp -草稿:排序搜索树和精炼游戏棋盘表示以改进草稿的多代理系统","authors":"V. Duarte, Rita Maria Silva Julia","doi":"10.1109/ICTAI.2012.159","DOIUrl":null,"url":null,"abstract":"In this paper the authors present an extension of the automatic player system MP-Draughts (MultiPhase-Draughts): a self-learning multi-agent environment for draughts composed of 26 Multiple Layer Perceptrons (MLPs). The weights of the MLPs are updated by Temporal Differences (TD). The search for the best move is conducted by a search algorithm based on Alpha-Beta pruning, Iterative Deepening and Table Transposition. One of the agents is trained in such a way that it becomes an expert in the initial stages of play and the remaining (25), in endgame stages. The endgame boards used to train the endgame agents are retrieved from an endgame board database and clustered by a Kohonem-SOM Neural Network (NN). The same Kohonem-SOM NN will also be used during the games to select which endgame agent is more suitable to play each time the endgame stage of play is reached. In this paper the authors propose the following modifications to improve the performance of MP-Draughts: first, to change the mapping of the board states such that, instead of indicating the presence or not of certain features, it indicates the number of elements pointed out by each feature, second, to order the search tree of each agent in such a way as to attenuate the innumerous re-evaluations of the same board state inherent to the iterative deepening strategy.","PeriodicalId":155588,"journal":{"name":"2012 IEEE 24th International Conference on Tools with Artificial Intelligence","volume":"4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2012-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":"{\"title\":\"MP-Draughts: Ordering the Search Tree and Refining the Game Board Representation to Improve a Multi-agent System for Draughts\",\"authors\":\"V. Duarte, Rita Maria Silva Julia\",\"doi\":\"10.1109/ICTAI.2012.159\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper the authors present an extension of the automatic player system MP-Draughts (MultiPhase-Draughts): a self-learning multi-agent environment for draughts composed of 26 Multiple Layer Perceptrons (MLPs). The weights of the MLPs are updated by Temporal Differences (TD). The search for the best move is conducted by a search algorithm based on Alpha-Beta pruning, Iterative Deepening and Table Transposition. One of the agents is trained in such a way that it becomes an expert in the initial stages of play and the remaining (25), in endgame stages. The endgame boards used to train the endgame agents are retrieved from an endgame board database and clustered by a Kohonem-SOM Neural Network (NN). The same Kohonem-SOM NN will also be used during the games to select which endgame agent is more suitable to play each time the endgame stage of play is reached. In this paper the authors propose the following modifications to improve the performance of MP-Draughts: first, to change the mapping of the board states such that, instead of indicating the presence or not of certain features, it indicates the number of elements pointed out by each feature, second, to order the search tree of each agent in such a way as to attenuate the innumerous re-evaluations of the same board state inherent to the iterative deepening strategy.\",\"PeriodicalId\":155588,\"journal\":{\"name\":\"2012 IEEE 24th International Conference on Tools with Artificial Intelligence\",\"volume\":\"4 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2012-11-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"6\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2012 IEEE 24th International Conference on Tools with Artificial Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICTAI.2012.159\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2012 IEEE 24th International Conference on Tools with Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICTAI.2012.159","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6

摘要

在本文中,作者提出了自动棋手系统mp - drafts (multiphase - drafts)的扩展:一个由26个多层感知机(mlp)组成的自学习多智能体环境。mlp的权重由时间差(Temporal difference, TD)更新。采用基于Alpha-Beta剪枝、迭代深化和表置换的搜索算法寻找最优走法。其中一个智能体被训练成在游戏的初始阶段成为专家,而剩下的(25个)则在游戏的最后阶段成为专家。用于训练终局代理的终局板从终局板数据库中检索,并通过Kohonem-SOM神经网络(NN)聚类。同样的Kohonem-SOM神经网络也将在游戏过程中使用,每次达到游戏的终局阶段时,选择哪个终局代理更适合玩。在本文中,作者提出了以下修改以提高mp -草稿的性能:首先,改变棋盘状态的映射,使其不再指示某些特征的存在与否,而是指示每个特征所指出的元素的数量;其次,对每个代理的搜索树进行排序,以减少迭代深化策略固有的对同一棋盘状态的无数次重新评估。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
MP-Draughts: Ordering the Search Tree and Refining the Game Board Representation to Improve a Multi-agent System for Draughts
In this paper the authors present an extension of the automatic player system MP-Draughts (MultiPhase-Draughts): a self-learning multi-agent environment for draughts composed of 26 Multiple Layer Perceptrons (MLPs). The weights of the MLPs are updated by Temporal Differences (TD). The search for the best move is conducted by a search algorithm based on Alpha-Beta pruning, Iterative Deepening and Table Transposition. One of the agents is trained in such a way that it becomes an expert in the initial stages of play and the remaining (25), in endgame stages. The endgame boards used to train the endgame agents are retrieved from an endgame board database and clustered by a Kohonem-SOM Neural Network (NN). The same Kohonem-SOM NN will also be used during the games to select which endgame agent is more suitable to play each time the endgame stage of play is reached. In this paper the authors propose the following modifications to improve the performance of MP-Draughts: first, to change the mapping of the board states such that, instead of indicating the presence or not of certain features, it indicates the number of elements pointed out by each feature, second, to order the search tree of each agent in such a way as to attenuate the innumerous re-evaluations of the same board state inherent to the iterative deepening strategy.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信