社会排斥是机器学习机制的副作用

A. Tertyshnikova, U. O. Pavlova, M. V. Cimbal
{"title":"社会排斥是机器学习机制的副作用","authors":"A. Tertyshnikova, U. O. Pavlova, M. V. Cimbal","doi":"10.26425/2658-347x-2022-5-4-23-30","DOIUrl":null,"url":null,"abstract":"The development of neural network technologies leads to their integration in decision-making processes at the level of such important social institutions as healthcare, education, employment, etc. This situation brings up the question of the correctness of artificial intelligence decisions and their consequences. The aim of this work is to consider the origin and replication of social exclusion, inequality and discrimination in society as a result of neurotraining. Neurotraining understood as the principles of any neural networks’ training. Social exclusion and the resulting discrimination in decisions made by artificial intelligence is considered as a consequence of the big data processing principles. The authors review the theories of foreign and Russian authors concerning the impact of artificial intelligence on strengthening the existing social order, as well as problems with processing and interpreting data for training computer systems on them. Real situations of the specifics of the data itself and its processing that have led to increased inequality and exclusion are also given. The conclusion about the sources of social exclusion and stigmatization in society is made due to the similarity between natural and artificial neural networks functioning. The authors suggest that it is the principles of neurotraining in a “natural” society that lead not only to discrimination at the macro level, but also cause vivid negative reactions towards representatives of the exclusive groups, for example, interethnic hatred, homophobia, sexism, etc. The question about the possibility of studying “natural” society in comparison with “artificial” one is raised.","PeriodicalId":52710,"journal":{"name":"Tsifrovaia sotsiologiia","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2023-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Social exclusion as a side effect of machine learning mechanisms\",\"authors\":\"A. Tertyshnikova, U. O. Pavlova, M. V. Cimbal\",\"doi\":\"10.26425/2658-347x-2022-5-4-23-30\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The development of neural network technologies leads to their integration in decision-making processes at the level of such important social institutions as healthcare, education, employment, etc. This situation brings up the question of the correctness of artificial intelligence decisions and their consequences. The aim of this work is to consider the origin and replication of social exclusion, inequality and discrimination in society as a result of neurotraining. Neurotraining understood as the principles of any neural networks’ training. Social exclusion and the resulting discrimination in decisions made by artificial intelligence is considered as a consequence of the big data processing principles. The authors review the theories of foreign and Russian authors concerning the impact of artificial intelligence on strengthening the existing social order, as well as problems with processing and interpreting data for training computer systems on them. Real situations of the specifics of the data itself and its processing that have led to increased inequality and exclusion are also given. The conclusion about the sources of social exclusion and stigmatization in society is made due to the similarity between natural and artificial neural networks functioning. The authors suggest that it is the principles of neurotraining in a “natural” society that lead not only to discrimination at the macro level, but also cause vivid negative reactions towards representatives of the exclusive groups, for example, interethnic hatred, homophobia, sexism, etc. The question about the possibility of studying “natural” society in comparison with “artificial” one is raised.\",\"PeriodicalId\":52710,\"journal\":{\"name\":\"Tsifrovaia sotsiologiia\",\"volume\":\" \",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-01-31\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Tsifrovaia sotsiologiia\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.26425/2658-347x-2022-5-4-23-30\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Tsifrovaia sotsiologiia","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.26425/2658-347x-2022-5-4-23-30","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

神经网络技术的发展导致它们在诸如医疗保健、教育、就业等重要社会机构层面的决策过程中得到整合。这种情况提出了人工智能决策的正确性及其后果的问题。这项工作的目的是考虑作为神经训练结果的社会排斥、不平等和歧视的起源和复制。神经训练被理解为任何神经网络训练的原则。人工智能在决策中产生的社会排斥和歧视被认为是大数据处理原则的结果。作者回顾了外国和俄罗斯作者关于人工智能对加强现有社会秩序的影响的理论,以及处理和解释用于训练计算机系统的数据的问题。还给出了导致不平等和排斥加剧的数据本身及其处理的具体情况。由于自然神经网络和人工神经网络功能的相似性,得出了社会排斥和社会污名化的来源。作者认为,“自然”社会中的神经训练原则不仅导致了宏观层面的歧视,而且导致了对排他性群体代表的生动的负面反应,例如种族间仇恨、同性恋恐惧症、性别歧视等。提出了研究“自然”社会与研究“人工”社会的可能性的问题。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Social exclusion as a side effect of machine learning mechanisms
The development of neural network technologies leads to their integration in decision-making processes at the level of such important social institutions as healthcare, education, employment, etc. This situation brings up the question of the correctness of artificial intelligence decisions and their consequences. The aim of this work is to consider the origin and replication of social exclusion, inequality and discrimination in society as a result of neurotraining. Neurotraining understood as the principles of any neural networks’ training. Social exclusion and the resulting discrimination in decisions made by artificial intelligence is considered as a consequence of the big data processing principles. The authors review the theories of foreign and Russian authors concerning the impact of artificial intelligence on strengthening the existing social order, as well as problems with processing and interpreting data for training computer systems on them. Real situations of the specifics of the data itself and its processing that have led to increased inequality and exclusion are also given. The conclusion about the sources of social exclusion and stigmatization in society is made due to the similarity between natural and artificial neural networks functioning. The authors suggest that it is the principles of neurotraining in a “natural” society that lead not only to discrimination at the macro level, but also cause vivid negative reactions towards representatives of the exclusive groups, for example, interethnic hatred, homophobia, sexism, etc. The question about the possibility of studying “natural” society in comparison with “artificial” one is raised.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
48
审稿时长
8 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信