Numeric vs. verbal information: The influence of information quantifiability in Human–AI vs. Human–Human decision support

Eileen Roesler , Tobias Rieger , Markus Langer
{"title":"Numeric vs. verbal information: The influence of information quantifiability in Human–AI vs. Human–Human decision support","authors":"Eileen Roesler ,&nbsp;Tobias Rieger ,&nbsp;Markus Langer","doi":"10.1016/j.chbah.2024.100116","DOIUrl":null,"url":null,"abstract":"<div><div>A number of factors, including different task characteristics, influence trust in human vs. AI decision support. In particular, the aspect of information quantifiability could influence trust and dependence, especially considering that human and AI support may have varying strengths in assessing criteria that differ in their quantifiability. To investigate the effect of information quantifiability we conducted an online experiment (<span><math><mrow><mi>N</mi><mo>=</mo><mn>204</mn></mrow></math></span>) with a 2 (support agent: AI vs. human) <span><math><mo>×</mo></math></span> 2 (quantifiability: low vs. high) between-subjects design, using a simulated recruitment task. The support agent was manipulated via framing, while quantifiability was manipulated by the evaluation criteria in the recruitment paradigm. The analysis revealed higher trust for human over AI support. Moreover, trust was higher in the low than in the high quantifiability condition. Counterintuitively, participants rated the applicants as less qualified than their support agent’s rating, especially noticeable in the low quantifiability condition. Besides reinforcing earlier findings showing higher trust towards human experts than towards AI and showcasing the importance of information quantifiability, the present study also raises questions concerning the perceived leniency of support agents and its impact on trust and behavior.</div></div>","PeriodicalId":100324,"journal":{"name":"Computers in Human Behavior: Artificial Humans","volume":"3 ","pages":"Article 100116"},"PeriodicalIF":0.0000,"publicationDate":"2024-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computers in Human Behavior: Artificial Humans","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2949882124000768","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

A number of factors, including different task characteristics, influence trust in human vs. AI decision support. In particular, the aspect of information quantifiability could influence trust and dependence, especially considering that human and AI support may have varying strengths in assessing criteria that differ in their quantifiability. To investigate the effect of information quantifiability we conducted an online experiment (N=204) with a 2 (support agent: AI vs. human) × 2 (quantifiability: low vs. high) between-subjects design, using a simulated recruitment task. The support agent was manipulated via framing, while quantifiability was manipulated by the evaluation criteria in the recruitment paradigm. The analysis revealed higher trust for human over AI support. Moreover, trust was higher in the low than in the high quantifiability condition. Counterintuitively, participants rated the applicants as less qualified than their support agent’s rating, especially noticeable in the low quantifiability condition. Besides reinforcing earlier findings showing higher trust towards human experts than towards AI and showcasing the importance of information quantifiability, the present study also raises questions concerning the perceived leniency of support agents and its impact on trust and behavior.
数字与语言信息:信息可量化在人-人工智能与人-人决策支持中的影响
许多因素,包括不同的任务特征,会影响对人类与人工智能决策支持的信任。特别是,信息可量化方面可能影响信任和依赖,特别是考虑到人类和人工智能支持在评估可量化性不同的标准方面可能具有不同的优势。为了研究信息可量化性的影响,我们进行了一项在线实验(N=204),采用2(支持代理:人工智能与人类)× 2(可量化性:低与高)的受试者间设计,使用模拟招聘任务。支持主体通过框架操纵,可量化性通过招募范式中的评价标准操纵。分析显示,与人工智能的支持相比,人们对人类的信任更高。此外,低量化条件下的信任度高于高量化条件下的信任度。与直觉相反,参与者对申请人的评价不如他们的支持代理人的评价,尤其是在低可量化条件下。除了加强早期的研究结果,即对人类专家的信任度高于对人工智能的信任度,并展示信息可量化的重要性外,本研究还提出了有关支持代理的感知宽容度及其对信任和行为的影响的问题。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信