Arpita Biswas, M. Kołczyńska, Saana Rantanen, Polina Rozenshtein
{"title":"群体内偏见和平衡数据的作用:人类和机器再犯风险预测的比较","authors":"Arpita Biswas, M. Kołczyńska, Saana Rantanen, Polina Rozenshtein","doi":"10.1145/3378393.3402507","DOIUrl":null,"url":null,"abstract":"Fairness and bias in automated decision-making gain importance as the prevalence of algorithms increases in different areas of social life. This paper contributes to the discussion of algorithmic fairness with a crowdsourced vignette survey on recidivism risk assessment, which we compare to previous studies on this topic and to predictions of an automated recidivism risk tool. We use the case of the Correctional Offender Management Profiling for Alternative Sanctions (COMPAS) and the Broward County dataset of pre-trial defendants as a data source and for purposes of comparability with the earlier analysis. In our survey, each respondent assessed recidivism risk for a set of vignettes describing real defendants, where each set was balanced with regard to the defendants' race and re-offender status. The survey ensured a 50: 50 ratio of black and white respondents. We found that predictions in our survey---while less accurate---were considerably more fair in terms of equalized odds than previous surveys. We attribute it to the differences in survey design: using the balanced set of vignettes and not providing feedback after responding to each vignette. We also analyzed the performance and fairness of predictions by race of respondent and defendant. We found that both white and black respondents tend to favor defendants of their own race, but the magnitude of the effect is relatively small. In addition to the survey, we train two statistical models, one trained with balanced data and other with unbalanced data. We observe that the model trained on balanced data is substantially more fair and possess less in-group bias.","PeriodicalId":176951,"journal":{"name":"Proceedings of the 3rd ACM SIGCAS Conference on Computing and Sustainable Societies","volume":"62 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":"{\"title\":\"The Role of In-Group Bias and Balanced Data: A Comparison of Human and Machine Recidivism Risk Predictions\",\"authors\":\"Arpita Biswas, M. Kołczyńska, Saana Rantanen, Polina Rozenshtein\",\"doi\":\"10.1145/3378393.3402507\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Fairness and bias in automated decision-making gain importance as the prevalence of algorithms increases in different areas of social life. This paper contributes to the discussion of algorithmic fairness with a crowdsourced vignette survey on recidivism risk assessment, which we compare to previous studies on this topic and to predictions of an automated recidivism risk tool. We use the case of the Correctional Offender Management Profiling for Alternative Sanctions (COMPAS) and the Broward County dataset of pre-trial defendants as a data source and for purposes of comparability with the earlier analysis. In our survey, each respondent assessed recidivism risk for a set of vignettes describing real defendants, where each set was balanced with regard to the defendants' race and re-offender status. The survey ensured a 50: 50 ratio of black and white respondents. We found that predictions in our survey---while less accurate---were considerably more fair in terms of equalized odds than previous surveys. We attribute it to the differences in survey design: using the balanced set of vignettes and not providing feedback after responding to each vignette. We also analyzed the performance and fairness of predictions by race of respondent and defendant. We found that both white and black respondents tend to favor defendants of their own race, but the magnitude of the effect is relatively small. In addition to the survey, we train two statistical models, one trained with balanced data and other with unbalanced data. We observe that the model trained on balanced data is substantially more fair and possess less in-group bias.\",\"PeriodicalId\":176951,\"journal\":{\"name\":\"Proceedings of the 3rd ACM SIGCAS Conference on Computing and Sustainable Societies\",\"volume\":\"62 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-06-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"8\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 3rd ACM SIGCAS Conference on Computing and Sustainable Societies\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3378393.3402507\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 3rd ACM SIGCAS Conference on Computing and Sustainable Societies","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3378393.3402507","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
The Role of In-Group Bias and Balanced Data: A Comparison of Human and Machine Recidivism Risk Predictions
Fairness and bias in automated decision-making gain importance as the prevalence of algorithms increases in different areas of social life. This paper contributes to the discussion of algorithmic fairness with a crowdsourced vignette survey on recidivism risk assessment, which we compare to previous studies on this topic and to predictions of an automated recidivism risk tool. We use the case of the Correctional Offender Management Profiling for Alternative Sanctions (COMPAS) and the Broward County dataset of pre-trial defendants as a data source and for purposes of comparability with the earlier analysis. In our survey, each respondent assessed recidivism risk for a set of vignettes describing real defendants, where each set was balanced with regard to the defendants' race and re-offender status. The survey ensured a 50: 50 ratio of black and white respondents. We found that predictions in our survey---while less accurate---were considerably more fair in terms of equalized odds than previous surveys. We attribute it to the differences in survey design: using the balanced set of vignettes and not providing feedback after responding to each vignette. We also analyzed the performance and fairness of predictions by race of respondent and defendant. We found that both white and black respondents tend to favor defendants of their own race, but the magnitude of the effect is relatively small. In addition to the survey, we train two statistical models, one trained with balanced data and other with unbalanced data. We observe that the model trained on balanced data is substantially more fair and possess less in-group bias.