{"title":"偏见产生偏见:有偏见的嵌入对扩散模型的影响","authors":"Sahil Kuchlous, Marvin Li, Jeffrey G. Wang","doi":"arxiv-2409.09569","DOIUrl":null,"url":null,"abstract":"With the growing adoption of Text-to-Image (TTI) systems, the social biases\nof these models have come under increased scrutiny. Herein we conduct a\nsystematic investigation of one such source of bias for diffusion models:\nembedding spaces. First, because traditional classifier-based fairness\ndefinitions require true labels not present in generative modeling, we propose\nstatistical group fairness criteria based on a model's internal representation\nof the world. Using these definitions, we demonstrate theoretically and\nempirically that an unbiased text embedding space for input prompts is a\nnecessary condition for representationally balanced diffusion models, meaning\nthe distribution of generated images satisfy diversity requirements with\nrespect to protected attributes. Next, we investigate the impact of biased\nembeddings on evaluating the alignment between generated images and prompts, a\nprocess which is commonly used to assess diffusion models. We find that biased\nmultimodal embeddings like CLIP can result in lower alignment scores for\nrepresentationally balanced TTI models, thus rewarding unfair behavior.\nFinally, we develop a theoretical framework through which biases in alignment\nevaluation can be studied and propose bias mitigation methods. By specifically\nadapting the perspective of embedding spaces, we establish new fairness\nconditions for diffusion model development and evaluation.","PeriodicalId":501112,"journal":{"name":"arXiv - CS - Computers and Society","volume":"118 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Bias Begets Bias: The Impact of Biased Embeddings on Diffusion Models\",\"authors\":\"Sahil Kuchlous, Marvin Li, Jeffrey G. Wang\",\"doi\":\"arxiv-2409.09569\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"With the growing adoption of Text-to-Image (TTI) systems, the social biases\\nof these models have come under increased scrutiny. Herein we conduct a\\nsystematic investigation of one such source of bias for diffusion models:\\nembedding spaces. First, because traditional classifier-based fairness\\ndefinitions require true labels not present in generative modeling, we propose\\nstatistical group fairness criteria based on a model's internal representation\\nof the world. Using these definitions, we demonstrate theoretically and\\nempirically that an unbiased text embedding space for input prompts is a\\nnecessary condition for representationally balanced diffusion models, meaning\\nthe distribution of generated images satisfy diversity requirements with\\nrespect to protected attributes. Next, we investigate the impact of biased\\nembeddings on evaluating the alignment between generated images and prompts, a\\nprocess which is commonly used to assess diffusion models. We find that biased\\nmultimodal embeddings like CLIP can result in lower alignment scores for\\nrepresentationally balanced TTI models, thus rewarding unfair behavior.\\nFinally, we develop a theoretical framework through which biases in alignment\\nevaluation can be studied and propose bias mitigation methods. By specifically\\nadapting the perspective of embedding spaces, we establish new fairness\\nconditions for diffusion model development and evaluation.\",\"PeriodicalId\":501112,\"journal\":{\"name\":\"arXiv - CS - Computers and Society\",\"volume\":\"118 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Computers and Society\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.09569\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computers and Society","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.09569","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Bias Begets Bias: The Impact of Biased Embeddings on Diffusion Models
With the growing adoption of Text-to-Image (TTI) systems, the social biases
of these models have come under increased scrutiny. Herein we conduct a
systematic investigation of one such source of bias for diffusion models:
embedding spaces. First, because traditional classifier-based fairness
definitions require true labels not present in generative modeling, we propose
statistical group fairness criteria based on a model's internal representation
of the world. Using these definitions, we demonstrate theoretically and
empirically that an unbiased text embedding space for input prompts is a
necessary condition for representationally balanced diffusion models, meaning
the distribution of generated images satisfy diversity requirements with
respect to protected attributes. Next, we investigate the impact of biased
embeddings on evaluating the alignment between generated images and prompts, a
process which is commonly used to assess diffusion models. We find that biased
multimodal embeddings like CLIP can result in lower alignment scores for
representationally balanced TTI models, thus rewarding unfair behavior.
Finally, we develop a theoretical framework through which biases in alignment
evaluation can be studied and propose bias mitigation methods. By specifically
adapting the perspective of embedding spaces, we establish new fairness
conditions for diffusion model development and evaluation.