SAF:利益相关者关于机器学习开发实践公平性的协议。

IF 2.7 2区 哲学 Q1 ENGINEERING, MULTIDISCIPLINARY
Georgina Curto, Flavio Comim
{"title":"SAF:利益相关者关于机器学习开发实践公平性的协议。","authors":"Georgina Curto,&nbsp;Flavio Comim","doi":"10.1007/s11948-023-00448-y","DOIUrl":null,"url":null,"abstract":"<p><p>This paper clarifies why bias cannot be completely mitigated in Machine Learning (ML) and proposes an end-to-end methodology to translate the ethical principle of justice and fairness into the practice of ML development as an ongoing agreement with stakeholders. The pro-ethical iterative process presented in the paper aims to challenge asymmetric power dynamics in the fairness decision making within ML design and support ML development teams to identify, mitigate and monitor bias at each step of ML systems development. The process also provides guidance on how to explain the always imperfect trade-offs in terms of bias to users.</p>","PeriodicalId":49564,"journal":{"name":"Science and Engineering Ethics","volume":"29 4","pages":"29"},"PeriodicalIF":2.7000,"publicationDate":"2023-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10366323/pdf/","citationCount":"0","resultStr":"{\"title\":\"SAF: Stakeholders' Agreement on Fairness in the Practice of Machine Learning Development.\",\"authors\":\"Georgina Curto,&nbsp;Flavio Comim\",\"doi\":\"10.1007/s11948-023-00448-y\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>This paper clarifies why bias cannot be completely mitigated in Machine Learning (ML) and proposes an end-to-end methodology to translate the ethical principle of justice and fairness into the practice of ML development as an ongoing agreement with stakeholders. The pro-ethical iterative process presented in the paper aims to challenge asymmetric power dynamics in the fairness decision making within ML design and support ML development teams to identify, mitigate and monitor bias at each step of ML systems development. The process also provides guidance on how to explain the always imperfect trade-offs in terms of bias to users.</p>\",\"PeriodicalId\":49564,\"journal\":{\"name\":\"Science and Engineering Ethics\",\"volume\":\"29 4\",\"pages\":\"29\"},\"PeriodicalIF\":2.7000,\"publicationDate\":\"2023-07-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10366323/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Science and Engineering Ethics\",\"FirstCategoryId\":\"98\",\"ListUrlMain\":\"https://doi.org/10.1007/s11948-023-00448-y\",\"RegionNum\":2,\"RegionCategory\":\"哲学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, MULTIDISCIPLINARY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Science and Engineering Ethics","FirstCategoryId":"98","ListUrlMain":"https://doi.org/10.1007/s11948-023-00448-y","RegionNum":2,"RegionCategory":"哲学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

摘要

本文阐明了为什么在机器学习(ML)中不能完全减轻偏见,并提出了一种端到端的方法,将正义和公平的道德原则转化为ML开发的实践,作为与利益相关者的持续协议。本文提出的亲道德迭代过程旨在挑战机器学习设计中公平决策中的不对称权力动态,并支持机器学习开发团队在机器学习系统开发的每个步骤中识别、减轻和监控偏见。该过程还提供了如何向用户解释总是不完美的权衡的指导。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

SAF: Stakeholders' Agreement on Fairness in the Practice of Machine Learning Development.

SAF: Stakeholders' Agreement on Fairness in the Practice of Machine Learning Development.

SAF: Stakeholders' Agreement on Fairness in the Practice of Machine Learning Development.

SAF: Stakeholders' Agreement on Fairness in the Practice of Machine Learning Development.

This paper clarifies why bias cannot be completely mitigated in Machine Learning (ML) and proposes an end-to-end methodology to translate the ethical principle of justice and fairness into the practice of ML development as an ongoing agreement with stakeholders. The pro-ethical iterative process presented in the paper aims to challenge asymmetric power dynamics in the fairness decision making within ML design and support ML development teams to identify, mitigate and monitor bias at each step of ML systems development. The process also provides guidance on how to explain the always imperfect trade-offs in terms of bias to users.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Science and Engineering Ethics
Science and Engineering Ethics 综合性期刊-工程:综合
CiteScore
10.70
自引率
5.40%
发文量
54
审稿时长
>12 weeks
期刊介绍: Science and Engineering Ethics is an international multidisciplinary journal dedicated to exploring ethical issues associated with science and engineering, covering professional education, research and practice as well as the effects of technological innovations and research findings on society. While the focus of this journal is on science and engineering, contributions from a broad range of disciplines, including social sciences and humanities, are welcomed. Areas of interest include, but are not limited to, ethics of new and emerging technologies, research ethics, computer ethics, energy ethics, animals and human subjects ethics, ethics education in science and engineering, ethics in design, biomedical ethics, values in technology and innovation. We welcome contributions that deal with these issues from an international perspective, particularly from countries that are underrepresented in these discussions.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信