为政府提供可解释的人工智能:在受影响者看来,解释的类型对算法决策的准确性、公平性和可信度有影响吗?

IF 7.8 1区 管理学 Q1 INFORMATION SCIENCE & LIBRARY SCIENCE
Naomi Aoki , Tomohiko Tatsumi , Go Naruse , Kentaro Maeda
{"title":"为政府提供可解释的人工智能:在受影响者看来,解释的类型对算法决策的准确性、公平性和可信度有影响吗?","authors":"Naomi Aoki ,&nbsp;Tomohiko Tatsumi ,&nbsp;Go Naruse ,&nbsp;Kentaro Maeda","doi":"10.1016/j.giq.2024.101965","DOIUrl":null,"url":null,"abstract":"<div><p>Amidst concerns over biased and misguided government decisions arrived at through algorithmic treatment, it is important for members of society to be able to perceive that public authorities are making fair, accurate, and trustworthy decisions. Inspired in part by equity and procedural justice theories and by theories of attitudes towards technologies, we posited that the perception of these attributes of decisions is influenced by the type of explanation offered, which can be input-based, group-based, case-based, or counterfactual. We tested our hypotheses with two studies, each of which involved a pre-registered online survey experiment conducted in December 2022. In both studies, the subjects (<em>N</em> = 1200) were officers in high positions at stock companies registered in Japan, who were presented with a scenario consisting of an algorithmic decision made by a public authority: a ministry's decision to reject a grant application from their company (Study 1) and a tax authority's decision to select their company for an on-site tax inspection (Study 2). The studies revealed that offering the subjects some type of explanation had a positive effect on their attitude towards a decision, to various extents, although the detailed results of the two studies are not robust. These findings call for a nuanced inquiry, both in research and practice, into how to best design explanations of algorithmic decisions from societal and human-centric perspectives in different decision-making contexts.</p></div>","PeriodicalId":48258,"journal":{"name":"Government Information Quarterly","volume":"41 4","pages":"Article 101965"},"PeriodicalIF":7.8000,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0740624X24000571/pdfft?md5=79e232415d01bd4e88037e2540f7bb9f&pid=1-s2.0-S0740624X24000571-main.pdf","citationCount":"0","resultStr":"{\"title\":\"Explainable AI for government: Does the type of explanation matter to the accuracy, fairness, and trustworthiness of an algorithmic decision as perceived by those who are affected?\",\"authors\":\"Naomi Aoki ,&nbsp;Tomohiko Tatsumi ,&nbsp;Go Naruse ,&nbsp;Kentaro Maeda\",\"doi\":\"10.1016/j.giq.2024.101965\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Amidst concerns over biased and misguided government decisions arrived at through algorithmic treatment, it is important for members of society to be able to perceive that public authorities are making fair, accurate, and trustworthy decisions. Inspired in part by equity and procedural justice theories and by theories of attitudes towards technologies, we posited that the perception of these attributes of decisions is influenced by the type of explanation offered, which can be input-based, group-based, case-based, or counterfactual. We tested our hypotheses with two studies, each of which involved a pre-registered online survey experiment conducted in December 2022. In both studies, the subjects (<em>N</em> = 1200) were officers in high positions at stock companies registered in Japan, who were presented with a scenario consisting of an algorithmic decision made by a public authority: a ministry's decision to reject a grant application from their company (Study 1) and a tax authority's decision to select their company for an on-site tax inspection (Study 2). The studies revealed that offering the subjects some type of explanation had a positive effect on their attitude towards a decision, to various extents, although the detailed results of the two studies are not robust. These findings call for a nuanced inquiry, both in research and practice, into how to best design explanations of algorithmic decisions from societal and human-centric perspectives in different decision-making contexts.</p></div>\",\"PeriodicalId\":48258,\"journal\":{\"name\":\"Government Information Quarterly\",\"volume\":\"41 4\",\"pages\":\"Article 101965\"},\"PeriodicalIF\":7.8000,\"publicationDate\":\"2024-09-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.sciencedirect.com/science/article/pii/S0740624X24000571/pdfft?md5=79e232415d01bd4e88037e2540f7bb9f&pid=1-s2.0-S0740624X24000571-main.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Government Information Quarterly\",\"FirstCategoryId\":\"91\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0740624X24000571\",\"RegionNum\":1,\"RegionCategory\":\"管理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"INFORMATION SCIENCE & LIBRARY SCIENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Government Information Quarterly","FirstCategoryId":"91","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0740624X24000571","RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"INFORMATION SCIENCE & LIBRARY SCIENCE","Score":null,"Total":0}
引用次数: 0

摘要

在人们担心政府通过算法处理做出的决策存在偏见和误导的同时,让社会成员感受到公共机构做出的决策是公平、准确和值得信赖的,这一点非常重要。受公平和程序正义理论以及对技术的态度理论的部分启发,我们假设,对决策的这些属性的感知会受到所提供的解释类型的影响,解释类型可以是基于输入的、基于群体的、基于案例的或基于反事实的。我们通过两项研究验证了我们的假设,每项研究都涉及一项于 2022 年 12 月进行的预先登记的在线调查实验。在这两项研究中,受试者(N = 1200)都是在日本注册的股份公司的高级职员,他们被展示给一个由公共机构做出的算法决策组成的场景:一个部委决定拒绝其公司的拨款申请(研究 1),以及一个税务机关决定选择其公司进行现场税务检查(研究 2)。这两项研究显示,向受试者提供某种解释在不同程度上对他们对决定的态度产生了积极影响,尽管这两项研究的详细结果并不可靠。这些发现要求我们在研究和实践中进行细致入微的探索,研究如何在不同的决策背景下,从社会和以人为本的角度出发,为算法决策提供最佳解释。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Explainable AI for government: Does the type of explanation matter to the accuracy, fairness, and trustworthiness of an algorithmic decision as perceived by those who are affected?

Amidst concerns over biased and misguided government decisions arrived at through algorithmic treatment, it is important for members of society to be able to perceive that public authorities are making fair, accurate, and trustworthy decisions. Inspired in part by equity and procedural justice theories and by theories of attitudes towards technologies, we posited that the perception of these attributes of decisions is influenced by the type of explanation offered, which can be input-based, group-based, case-based, or counterfactual. We tested our hypotheses with two studies, each of which involved a pre-registered online survey experiment conducted in December 2022. In both studies, the subjects (N = 1200) were officers in high positions at stock companies registered in Japan, who were presented with a scenario consisting of an algorithmic decision made by a public authority: a ministry's decision to reject a grant application from their company (Study 1) and a tax authority's decision to select their company for an on-site tax inspection (Study 2). The studies revealed that offering the subjects some type of explanation had a positive effect on their attitude towards a decision, to various extents, although the detailed results of the two studies are not robust. These findings call for a nuanced inquiry, both in research and practice, into how to best design explanations of algorithmic decisions from societal and human-centric perspectives in different decision-making contexts.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Government Information Quarterly
Government Information Quarterly INFORMATION SCIENCE & LIBRARY SCIENCE-
CiteScore
15.70
自引率
16.70%
发文量
106
期刊介绍: Government Information Quarterly (GIQ) delves into the convergence of policy, information technology, government, and the public. It explores the impact of policies on government information flows, the role of technology in innovative government services, and the dynamic between citizens and governing bodies in the digital age. GIQ serves as a premier journal, disseminating high-quality research and insights that bridge the realms of policy, information technology, government, and public engagement.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信