真实与合成交叉光谱图像的人脸生物特征公平性评价

K. Lai, V. Shmerko, S. Yanushkevich
{"title":"真实与合成交叉光谱图像的人脸生物特征公平性评价","authors":"K. Lai, V. Shmerko, S. Yanushkevich","doi":"10.1109/CogMI56440.2022.00024","DOIUrl":null,"url":null,"abstract":"In this paper, we compare the performance and fairness metrics on visual and thermal images of faces, including the synthetic images of human subjects with face masks. The comparative experiment is performed on two datasets: the SpeakingFace and Thermal-Mask dataset. We assess fairness on real images and show how the same process can be applied to synthetic images. The chosen fairness metrics include demographic parity difference and equalized odds difference. While the demographic parity difference is assessed as 1.24 for random guessing in the process of face identification, it reaches 5.0 when both the precision and recall rate approach 99.99%. These results confirm that inherently biased datasets significantly impact the fairness of any biometric system. For biometric-enabled systems, fairness is related to the adequacy of the data to represent different groups of human subjects. In this paper, we focus on three demographic groups: age, gender, and ethnicity. A primary cause of biases with respect to these groups is the class imbalance introduced through the data collection process. To address the imbalanced datasets, the classes with fewer samples can be augmented with synthetic images to generate a more balanced dataset, resulting in less bias when training a machine learning system. The study shows that fairness is correlated to the performance of the system rather than to the genesis of the images (real or synthetic). The experiment on a simple 3-Block CNN with a precision and recall rate of 99.99% using the demographic parity difference as an estimate of fairness showed that among gender, ethnicity, and age, the latter is an attribute that is the most sensitive while age is the least one.","PeriodicalId":211430,"journal":{"name":"2022 IEEE 4th International Conference on Cognitive Machine Intelligence (CogMI)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Face Biometric Fairness Evaluation on Real vs Synthetic Cross-Spectral Images\",\"authors\":\"K. Lai, V. Shmerko, S. Yanushkevich\",\"doi\":\"10.1109/CogMI56440.2022.00024\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, we compare the performance and fairness metrics on visual and thermal images of faces, including the synthetic images of human subjects with face masks. The comparative experiment is performed on two datasets: the SpeakingFace and Thermal-Mask dataset. We assess fairness on real images and show how the same process can be applied to synthetic images. The chosen fairness metrics include demographic parity difference and equalized odds difference. While the demographic parity difference is assessed as 1.24 for random guessing in the process of face identification, it reaches 5.0 when both the precision and recall rate approach 99.99%. These results confirm that inherently biased datasets significantly impact the fairness of any biometric system. For biometric-enabled systems, fairness is related to the adequacy of the data to represent different groups of human subjects. In this paper, we focus on three demographic groups: age, gender, and ethnicity. A primary cause of biases with respect to these groups is the class imbalance introduced through the data collection process. To address the imbalanced datasets, the classes with fewer samples can be augmented with synthetic images to generate a more balanced dataset, resulting in less bias when training a machine learning system. The study shows that fairness is correlated to the performance of the system rather than to the genesis of the images (real or synthetic). The experiment on a simple 3-Block CNN with a precision and recall rate of 99.99% using the demographic parity difference as an estimate of fairness showed that among gender, ethnicity, and age, the latter is an attribute that is the most sensitive while age is the least one.\",\"PeriodicalId\":211430,\"journal\":{\"name\":\"2022 IEEE 4th International Conference on Cognitive Machine Intelligence (CogMI)\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-09-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE 4th International Conference on Cognitive Machine Intelligence (CogMI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CogMI56440.2022.00024\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 4th International Conference on Cognitive Machine Intelligence (CogMI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CogMI56440.2022.00024","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

在本文中,我们比较了人脸视觉图像和热图像的性能和公平性指标,包括人类受试者戴口罩的合成图像。在SpeakingFace和Thermal-Mask两个数据集上进行了对比实验。我们评估了真实图像的公平性,并展示了如何将相同的过程应用于合成图像。所选择的公平指标包括人口均等差异和均等赔率差异。在人脸识别过程中,随机猜测的人口统计等值差为1.24,当准确率和召回率均接近99.99%时,人口统计等值差达到5.0。这些结果证实,固有偏见的数据集显著影响任何生物识别系统的公平性。对于支持生物识别的系统,公平性与代表不同人类受试者群体的数据的充分性有关。在本文中,我们关注三个人口群体:年龄、性别和种族。对这些群体产生偏见的主要原因是通过数据收集过程引入的阶级不平衡。为了解决不平衡的数据集,可以用合成图像增强样本较少的类,以生成更平衡的数据集,从而在训练机器学习系统时减少偏差。研究表明,公平性与系统的性能相关,而与图像的起源(真实的或合成的)无关。在一个准确率和召回率为99.99%的简单3块CNN上进行的实验表明,在性别、种族和年龄中,后者是最敏感的属性,而年龄是最不敏感的属性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Face Biometric Fairness Evaluation on Real vs Synthetic Cross-Spectral Images
In this paper, we compare the performance and fairness metrics on visual and thermal images of faces, including the synthetic images of human subjects with face masks. The comparative experiment is performed on two datasets: the SpeakingFace and Thermal-Mask dataset. We assess fairness on real images and show how the same process can be applied to synthetic images. The chosen fairness metrics include demographic parity difference and equalized odds difference. While the demographic parity difference is assessed as 1.24 for random guessing in the process of face identification, it reaches 5.0 when both the precision and recall rate approach 99.99%. These results confirm that inherently biased datasets significantly impact the fairness of any biometric system. For biometric-enabled systems, fairness is related to the adequacy of the data to represent different groups of human subjects. In this paper, we focus on three demographic groups: age, gender, and ethnicity. A primary cause of biases with respect to these groups is the class imbalance introduced through the data collection process. To address the imbalanced datasets, the classes with fewer samples can be augmented with synthetic images to generate a more balanced dataset, resulting in less bias when training a machine learning system. The study shows that fairness is correlated to the performance of the system rather than to the genesis of the images (real or synthetic). The experiment on a simple 3-Block CNN with a precision and recall rate of 99.99% using the demographic parity difference as an estimate of fairness showed that among gender, ethnicity, and age, the latter is an attribute that is the most sensitive while age is the least one.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信