An Agent Bilateral Multi-issue Alternate Bidding Negotiation Protocol Based on Reinforcement Learning and its Application in E-commerce

Li Jian
{"title":"An Agent Bilateral Multi-issue Alternate Bidding Negotiation Protocol Based on Reinforcement Learning and its Application in E-commerce","authors":"Li Jian","doi":"10.1109/ISECS.2008.102","DOIUrl":null,"url":null,"abstract":"With the rapid development of multi-agent based E-commerce systems, on-line automatic negotiation protocol is often needed. But because of incomplete information agents have, the efficiency of on-line negotiation protocol is rather low. To overcome the problem, an on-line agent bilateral multi-issue alternate bidding negotiation protocol based on reinforcement learning is present. The reinforcement learning algorithm is presented to on-line learn the incomplete information of negotiation agent to enhance the efficiency of negotiation protocol. The protocol is applied to on-line multi-agent based electronic commerce. In the protocol experiment, three kinds of agents are used to compare with, which are no-learning agents (NA), static learning agents (SA) and dynamic learning agent (DA) in this paper. In static learning agent, the learning rate of Q-learning is set to 0.1 unchangeable, so itpsilas called static learning. While in dynamic learning proposed in this paper, the learning rate of Q-learning can change dynamically, so itpsilas called dynamic learning. Experiments show that the protocol present in this paper can help agents to negotiate more efficiently.","PeriodicalId":144075,"journal":{"name":"2008 International Symposium on Electronic Commerce and Security","volume":"78 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2008-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"11","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2008 International Symposium on Electronic Commerce and Security","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISECS.2008.102","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 11

Abstract

With the rapid development of multi-agent based E-commerce systems, on-line automatic negotiation protocol is often needed. But because of incomplete information agents have, the efficiency of on-line negotiation protocol is rather low. To overcome the problem, an on-line agent bilateral multi-issue alternate bidding negotiation protocol based on reinforcement learning is present. The reinforcement learning algorithm is presented to on-line learn the incomplete information of negotiation agent to enhance the efficiency of negotiation protocol. The protocol is applied to on-line multi-agent based electronic commerce. In the protocol experiment, three kinds of agents are used to compare with, which are no-learning agents (NA), static learning agents (SA) and dynamic learning agent (DA) in this paper. In static learning agent, the learning rate of Q-learning is set to 0.1 unchangeable, so itpsilas called static learning. While in dynamic learning proposed in this paper, the learning rate of Q-learning can change dynamically, so itpsilas called dynamic learning. Experiments show that the protocol present in this paper can help agents to negotiate more efficiently.
基于强化学习的Agent双边多议题竞价协商协议及其在电子商务中的应用
随着多智能体电子商务系统的快速发展,往往需要在线自动协商协议。但是由于agent所具有的信息不完全,使得在线协商协议的效率很低。为了解决这一问题,提出了一种基于强化学习的在线代理双边多议题备用投标协商协议。为了提高协商协议的效率,提出了在线学习协商代理不完全信息的强化学习算法。该协议应用于基于多代理的在线电子商务中。在协议实验中,本文采用无学习智能体(NA)、静态学习智能体(SA)和动态学习智能体(DA)三种智能体进行对比。在静态学习代理中,Q-learning的学习率被设置为0.1不变,因此称为静态学习。而在本文提出的动态学习中,q学习的学习率是动态变化的,因此称为动态学习。实验表明,本文提出的协议可以帮助agent更有效地进行协商。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信