{"title":"Decentralized Dynamics and Fast Convergence in the Assignment Game: Extended Abstract","authors":"Bary S. R. Pradelski","doi":"10.1145/2764468.2764470","DOIUrl":null,"url":null,"abstract":"We study decentralized learning dynamics for the classic assignment game with transferable utility [Shapley and Shubik 1972]. In our model agents follow an aspiration adjustment process based on their experienced payoffs (see [Sauermann and Selten 1962], [Nax and Pradelski 2014]). At random points in time firms and workers match, break up, and re-match in the search for better opportunities. Agents have aspiration levels that they adjust based on their experienced payoffs. When matched an agent occasionally tries to succeed with a higher bid than his current aspiration level. When single an agent lowers his aspiration level in the hope of attracting a partner. In particular agents have no knowledge about other players' payoffs or actions and they update their behavior in a myopic fashion. Behavior fluctuates according to a random variable that reflects current market sentiment: sometimes the firms exhibit greater price stickiness than the workers, and at other times the reverse holds. We show that this stochastic learning process converges in polynomial time to the core. While convergence to the core is known for some types of decentralized dynamics this paper is the first to prove {polynomial time convergence}, a crucial feature from an explanatory and market design standpoint. We also show that without market sentiment the dynamic exhibits exponential time convergence. The proof relies on novel results for random walks on graphs, and more generally suggests a fruitful connection between the theory of random walks and matching theory.","PeriodicalId":376992,"journal":{"name":"Proceedings of the Sixteenth ACM Conference on Economics and Computation","volume":"88 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2015-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"14","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the Sixteenth ACM Conference on Economics and Computation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/2764468.2764470","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 14
Abstract
We study decentralized learning dynamics for the classic assignment game with transferable utility [Shapley and Shubik 1972]. In our model agents follow an aspiration adjustment process based on their experienced payoffs (see [Sauermann and Selten 1962], [Nax and Pradelski 2014]). At random points in time firms and workers match, break up, and re-match in the search for better opportunities. Agents have aspiration levels that they adjust based on their experienced payoffs. When matched an agent occasionally tries to succeed with a higher bid than his current aspiration level. When single an agent lowers his aspiration level in the hope of attracting a partner. In particular agents have no knowledge about other players' payoffs or actions and they update their behavior in a myopic fashion. Behavior fluctuates according to a random variable that reflects current market sentiment: sometimes the firms exhibit greater price stickiness than the workers, and at other times the reverse holds. We show that this stochastic learning process converges in polynomial time to the core. While convergence to the core is known for some types of decentralized dynamics this paper is the first to prove {polynomial time convergence}, a crucial feature from an explanatory and market design standpoint. We also show that without market sentiment the dynamic exhibits exponential time convergence. The proof relies on novel results for random walks on graphs, and more generally suggests a fruitful connection between the theory of random walks and matching theory.
我们研究了具有可转移效用的经典分配博弈的分散学习动力学[Shapley and Shubik 1972]。在我们的模型中,代理人根据他们的经验收益遵循一个期望调整过程(参见[Sauermann and Selten 1962], [Nax and Pradelski 2014])。在寻找更好的机会的过程中,企业和工人在随机的时间点上配对、分手、再配对。代理人有期望水平,他们会根据自己的经验回报来调整。当匹配时,代理偶尔会尝试以高于其当前期望水平的出价获得成功。单身时,经纪人会降低自己的期望值,希望能吸引到合作伙伴。特别是,代理人不知道其他玩家的收益或行动,他们以短视的方式更新自己的行为。行为根据反映当前市场情绪的随机变量波动:有时公司比工人表现出更大的价格粘性,有时则相反。我们证明了这种随机学习过程在多项式时间内收敛到核心。虽然对于某些类型的分散动态来说,向核心收敛是已知的,但本文是第一个证明{多项式时间收敛}的论文,从解释和市场设计的角度来看,这是一个关键特征。我们还表明,在没有市场情绪的情况下,动态表现为指数时间收敛。该证明依赖于图上随机漫步的新结果,并且更普遍地表明随机漫步理论与匹配理论之间存在富有成效的联系。