Hedin Beattie, Lanier A Watkins, William H. Robinson, A. Rubin, Shari Watkins
{"title":"Measuring and Mitigating Bias in AI-Chatbots","authors":"Hedin Beattie, Lanier A Watkins, William H. Robinson, A. Rubin, Shari Watkins","doi":"10.1109/ICAA52185.2022.00023","DOIUrl":null,"url":null,"abstract":"The use of artificial intelligence (AI) to train conversational chatbots in the nuances of human interactions raises the concern of whether chatbots will demonstrate prejudice similar to that of humans, and thus require bias training. Ideally, a chatbot is void of racism, sexism, or any other offensive speech, however several well-known public instances indicate otherwise (e.g., Microsoft Taybot).In this paper, we explore the mechanisms of how open source conversational chatbots can learn bias, and we investigate potential solutions to mitigate this learned bias. To this end, we developed the Chatbot Bias Assessment Framework to measure bias in conversational chatbots, and then we devised an approach based on counter-stereotypic imagining to reduce this bias. This approach is non-intrusive to the chatbot, since it does not require altering any AI code or deleting any data from the original training dataset.","PeriodicalId":206047,"journal":{"name":"2022 IEEE International Conference on Assured Autonomy (ICAA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE International Conference on Assured Autonomy (ICAA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAA52185.2022.00023","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
The use of artificial intelligence (AI) to train conversational chatbots in the nuances of human interactions raises the concern of whether chatbots will demonstrate prejudice similar to that of humans, and thus require bias training. Ideally, a chatbot is void of racism, sexism, or any other offensive speech, however several well-known public instances indicate otherwise (e.g., Microsoft Taybot).In this paper, we explore the mechanisms of how open source conversational chatbots can learn bias, and we investigate potential solutions to mitigate this learned bias. To this end, we developed the Chatbot Bias Assessment Framework to measure bias in conversational chatbots, and then we devised an approach based on counter-stereotypic imagining to reduce this bias. This approach is non-intrusive to the chatbot, since it does not require altering any AI code or deleting any data from the original training dataset.