Ludovica Ilari, Giulia Rafaiani, M. Baldi, B. Giovanola
{"title":"基于机器学习的互联网通信过滤中的伦理偏见","authors":"Ludovica Ilari, Giulia Rafaiani, M. Baldi, B. Giovanola","doi":"10.1109/ETHICS57328.2023.10154975","DOIUrl":null,"url":null,"abstract":"The use of automated systems based on artificial intelligence and machine learning for filtering and moderating online communications has become commonplace. While this allows for high levels of efficiency and fine-grained control of malicious behaviors, it could also produce unintended disparities in treatment of legitimate users. In this paper, we aim at identifying some possible field-related biases in the wellknown Google Perspective API machine learning-based engine for controlling Internet communications. For this purpose, we consider communications in the fields of health, trade, finance, and defense and build a data set collecting Twitter-based online communications of the World Health Organization (WHO), World Trade Organization (WTO), International Monetary Fund (IMF) and North Atlantic Treaty Organization (NATO). Collected data are then analyzed through Perspective API to assign them an alleged likelihood of being abusive for specific emotional concepts, referred to as attributes. Upon analysis, discrimination between the considered users is identified for all attributes. This result, although preliminary, apparently indicates that Perspective API creates discrimination for field-related content as a result of semantic biases in the data, thus highlighting the need for an ethically sound design of these systems, following an ethics by design approach.","PeriodicalId":203527,"journal":{"name":"2023 IEEE International Symposium on Ethics in Engineering, Science, and Technology (ETHICS)","volume":"2006 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Ethical Biases in Machine Learning-based Filtering of Internet Communications\",\"authors\":\"Ludovica Ilari, Giulia Rafaiani, M. Baldi, B. Giovanola\",\"doi\":\"10.1109/ETHICS57328.2023.10154975\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The use of automated systems based on artificial intelligence and machine learning for filtering and moderating online communications has become commonplace. While this allows for high levels of efficiency and fine-grained control of malicious behaviors, it could also produce unintended disparities in treatment of legitimate users. In this paper, we aim at identifying some possible field-related biases in the wellknown Google Perspective API machine learning-based engine for controlling Internet communications. For this purpose, we consider communications in the fields of health, trade, finance, and defense and build a data set collecting Twitter-based online communications of the World Health Organization (WHO), World Trade Organization (WTO), International Monetary Fund (IMF) and North Atlantic Treaty Organization (NATO). Collected data are then analyzed through Perspective API to assign them an alleged likelihood of being abusive for specific emotional concepts, referred to as attributes. Upon analysis, discrimination between the considered users is identified for all attributes. This result, although preliminary, apparently indicates that Perspective API creates discrimination for field-related content as a result of semantic biases in the data, thus highlighting the need for an ethically sound design of these systems, following an ethics by design approach.\",\"PeriodicalId\":203527,\"journal\":{\"name\":\"2023 IEEE International Symposium on Ethics in Engineering, Science, and Technology (ETHICS)\",\"volume\":\"2006 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-05-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 IEEE International Symposium on Ethics in Engineering, Science, and Technology (ETHICS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ETHICS57328.2023.10154975\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE International Symposium on Ethics in Engineering, Science, and Technology (ETHICS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ETHICS57328.2023.10154975","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Ethical Biases in Machine Learning-based Filtering of Internet Communications
The use of automated systems based on artificial intelligence and machine learning for filtering and moderating online communications has become commonplace. While this allows for high levels of efficiency and fine-grained control of malicious behaviors, it could also produce unintended disparities in treatment of legitimate users. In this paper, we aim at identifying some possible field-related biases in the wellknown Google Perspective API machine learning-based engine for controlling Internet communications. For this purpose, we consider communications in the fields of health, trade, finance, and defense and build a data set collecting Twitter-based online communications of the World Health Organization (WHO), World Trade Organization (WTO), International Monetary Fund (IMF) and North Atlantic Treaty Organization (NATO). Collected data are then analyzed through Perspective API to assign them an alleged likelihood of being abusive for specific emotional concepts, referred to as attributes. Upon analysis, discrimination between the considered users is identified for all attributes. This result, although preliminary, apparently indicates that Perspective API creates discrimination for field-related content as a result of semantic biases in the data, thus highlighting the need for an ethically sound design of these systems, following an ethics by design approach.