When artificial minds negotiate: Dark personality and the Ultimatum Game in large language models

Vinícius Ferraz , Tamas Olah , Ratin Sazedul , Robert Schmidt , Christiane Schwieren
{"title":"When artificial minds negotiate: Dark personality and the Ultimatum Game in large language models","authors":"Vinícius Ferraz ,&nbsp;Tamas Olah ,&nbsp;Ratin Sazedul ,&nbsp;Robert Schmidt ,&nbsp;Christiane Schwieren","doi":"10.1016/j.chbah.2026.100281","DOIUrl":null,"url":null,"abstract":"<div><div>Personality prompts reshape how Large Language Models propose offers in economic games—but not how they respond to them. We show this by assigning graded Dark Factor of Personality profiles to 17 LLMs in the Ultimatum Game and benchmarking their decisions against human data. As proposers, LLMs shifted from 91% fair offers at the lowest selfishness level to 17% at the highest, closely tracking human patterns but with steeper gradients. As responders, no such shift occurred: acceptance rates remained uniformly high (<span><math><mo>∼</mo></math></span>80%) regardless of personality, failing to reproduce the punishment dynamics observed in humans. This asymmetry is theoretically informative. When incentive structures are explicit, personality and framing effects are attenuated—and proposing an offer is inherently more ambiguous than responding to one. Most strikingly, personality prompts changed what responders <em>articulated</em> but not how they <em>chose</em>: model justifications showed systematic shifts in fairness language, yet behavioral output remained flat. This dissociation between stated reasoning and revealed behavior indicates that LLMs achieve linguistic compliance with personality prompts without corresponding motivational change—approximating human strategic behavior only where surface-level heuristics suffice.</div></div>","PeriodicalId":100324,"journal":{"name":"Computers in Human Behavior: Artificial Humans","volume":"7 ","pages":"Article 100281"},"PeriodicalIF":0.0000,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computers in Human Behavior: Artificial Humans","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2949882126000320","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2026/2/24 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Personality prompts reshape how Large Language Models propose offers in economic games—but not how they respond to them. We show this by assigning graded Dark Factor of Personality profiles to 17 LLMs in the Ultimatum Game and benchmarking their decisions against human data. As proposers, LLMs shifted from 91% fair offers at the lowest selfishness level to 17% at the highest, closely tracking human patterns but with steeper gradients. As responders, no such shift occurred: acceptance rates remained uniformly high (80%) regardless of personality, failing to reproduce the punishment dynamics observed in humans. This asymmetry is theoretically informative. When incentive structures are explicit, personality and framing effects are attenuated—and proposing an offer is inherently more ambiguous than responding to one. Most strikingly, personality prompts changed what responders articulated but not how they chose: model justifications showed systematic shifts in fairness language, yet behavioral output remained flat. This dissociation between stated reasoning and revealed behavior indicates that LLMs achieve linguistic compliance with personality prompts without corresponding motivational change—approximating human strategic behavior only where surface-level heuristics suffice.
当人工思维谈判:大型语言模型中的黑暗人格和最后通牒游戏
个性会重塑大语言模型在经济博弈中提出报价的方式,但不会影响它们对报价的反应。我们通过在最后通牒游戏中分配17位法学硕士的人格黑暗因素档案,并根据人类数据对他们的决策进行基准测试,来证明这一点。作为提议者,法学硕士的公平出价从最低自私水平的91%上升到最高自私水平的17%,与人类模式密切相关,但梯度更大。作为应答者,没有发生这样的转变:无论性格如何,接受率都保持一致的高(约80%),未能重现在人类中观察到的惩罚动态。这种不对称在理论上提供了信息。当激励结构明确时,人格和框架效应就会减弱——提出提议本身就比回应提议更模棱两可。最引人注目的是,性格提示改变了应答者所表达的内容,但没有改变他们的选择方式:模型辩护显示出公平语言的系统性转变,但行为输出保持不变。陈述推理和揭示行为之间的这种分离表明,llm在没有相应动机改变的情况下实现了对人格提示的语言遵从——只有在表面层次的启发式足够的情况下才近似于人类的策略行为。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信
小红书