{"title":"ML-EAT:可解释和透明社会科学的多层次嵌入关联测试","authors":"Robert Wolfe, Alexis Hiniker, Bill Howe","doi":"arxiv-2408.01966","DOIUrl":null,"url":null,"abstract":"This research introduces the Multilevel Embedding Association Test (ML-EAT),\na method designed for interpretable and transparent measurement of intrinsic\nbias in language technologies. The ML-EAT addresses issues of ambiguity and\ndifficulty in interpreting the traditional EAT measurement by quantifying bias\nat three levels of increasing granularity: the differential association between\ntwo target concepts with two attribute concepts; the individual effect size of\neach target concept with two attribute concepts; and the association between\neach individual target concept and each individual attribute concept. Using the\nML-EAT, this research defines a taxonomy of EAT patterns describing the nine\npossible outcomes of an embedding association test, each of which is associated\nwith a unique EAT-Map, a novel four-quadrant visualization for interpreting the\nML-EAT. Empirical analysis of static and diachronic word embeddings, GPT-2\nlanguage models, and a CLIP language-and-image model shows that EAT patterns\nadd otherwise unobservable information about the component biases that make up\nan EAT; reveal the effects of prompting in zero-shot models; and can also\nidentify situations when cosine similarity is an ineffective metric, rendering\nan EAT unreliable. Our work contributes a method for rendering bias more\nobservable and interpretable, improving the transparency of computational\ninvestigations into human minds and societies.","PeriodicalId":501112,"journal":{"name":"arXiv - CS - Computers and Society","volume":"1 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"ML-EAT: A Multilevel Embedding Association Test for Interpretable and Transparent Social Science\",\"authors\":\"Robert Wolfe, Alexis Hiniker, Bill Howe\",\"doi\":\"arxiv-2408.01966\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This research introduces the Multilevel Embedding Association Test (ML-EAT),\\na method designed for interpretable and transparent measurement of intrinsic\\nbias in language technologies. The ML-EAT addresses issues of ambiguity and\\ndifficulty in interpreting the traditional EAT measurement by quantifying bias\\nat three levels of increasing granularity: the differential association between\\ntwo target concepts with two attribute concepts; the individual effect size of\\neach target concept with two attribute concepts; and the association between\\neach individual target concept and each individual attribute concept. Using the\\nML-EAT, this research defines a taxonomy of EAT patterns describing the nine\\npossible outcomes of an embedding association test, each of which is associated\\nwith a unique EAT-Map, a novel four-quadrant visualization for interpreting the\\nML-EAT. Empirical analysis of static and diachronic word embeddings, GPT-2\\nlanguage models, and a CLIP language-and-image model shows that EAT patterns\\nadd otherwise unobservable information about the component biases that make up\\nan EAT; reveal the effects of prompting in zero-shot models; and can also\\nidentify situations when cosine similarity is an ineffective metric, rendering\\nan EAT unreliable. Our work contributes a method for rendering bias more\\nobservable and interpretable, improving the transparency of computational\\ninvestigations into human minds and societies.\",\"PeriodicalId\":501112,\"journal\":{\"name\":\"arXiv - CS - Computers and Society\",\"volume\":\"1 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-08-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Computers and Society\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2408.01966\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computers and Society","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.01966","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
ML-EAT: A Multilevel Embedding Association Test for Interpretable and Transparent Social Science
This research introduces the Multilevel Embedding Association Test (ML-EAT),
a method designed for interpretable and transparent measurement of intrinsic
bias in language technologies. The ML-EAT addresses issues of ambiguity and
difficulty in interpreting the traditional EAT measurement by quantifying bias
at three levels of increasing granularity: the differential association between
two target concepts with two attribute concepts; the individual effect size of
each target concept with two attribute concepts; and the association between
each individual target concept and each individual attribute concept. Using the
ML-EAT, this research defines a taxonomy of EAT patterns describing the nine
possible outcomes of an embedding association test, each of which is associated
with a unique EAT-Map, a novel four-quadrant visualization for interpreting the
ML-EAT. Empirical analysis of static and diachronic word embeddings, GPT-2
language models, and a CLIP language-and-image model shows that EAT patterns
add otherwise unobservable information about the component biases that make up
an EAT; reveal the effects of prompting in zero-shot models; and can also
identify situations when cosine similarity is an ineffective metric, rendering
an EAT unreliable. Our work contributes a method for rendering bias more
observable and interpretable, improving the transparency of computational
investigations into human minds and societies.