Ariana L. Shaari, Anthony M. Saad, Aman M. Patel, Andrey Filimonov
{"title":"人工智能文本-图像平台在耳鼻喉科的人口统计学表征","authors":"Ariana L. Shaari, Anthony M. Saad, Aman M. Patel, Andrey Filimonov","doi":"10.1002/lio2.70152","DOIUrl":null,"url":null,"abstract":"<div>\n \n \n <section>\n \n <h3> Objective</h3>\n \n <p>Artificial intelligence (AI) text-to-image generators have a propensity to reflect stereotypes. This study investigates the perception of race and gender of AI-generated portraits of otolaryngologists, evaluating their accuracy against workforce demographics and whether they amplify existing social biases.</p>\n </section>\n \n <section>\n \n <h3> Methods</h3>\n \n <p>Three text-to-image platforms (DALL-E3, Runway, Midjourney) were prompted to generate portrait photos of otolaryngologists based on 29 categories, including personality traits, fellowship, and academic rank. 580 portrait photos were made per platform. Two reviewers characterized the gender and race of the 1740 portraits. Statistical analysis compared the demographics of AI outputs to existing demographic information.</p>\n </section>\n \n <section>\n \n <h3> Results</h3>\n \n <p>Of the 1740 AI-generated otolaryngologists generated, 88% of images were labeled as White, 4% Black, 6% Asian, 2% Indeterminate/Other race, 88% male, and 12% female. Across academic rank, the representation of White individuals was 97% (department chairs), 90% (program directors), 93% (professors), and 78% (residents). Male representation ranged from 90% (department chairs), 75% (program directors), 100% (professors), and 87% (residents). Runway produced more images of male (89% vs. 88% vs. 85%, <i>p</i> = 0.043) and White (92% vs. 88% vs. 80%, <i>p</i> < 0.001) otolaryngologists than DALL-E3 and Midjourney, respectively.</p>\n </section>\n \n <section>\n \n <h3> Conclusion</h3>\n \n <p>Text-to-image platforms demonstrated racial and gender biases, with notable differences compared to actual demographics. These platforms often underrepresented females and racial minority groups and overrepresented White males. These disparities underscore the need for the awareness of biases in AI, especially as these tools become more integrated into patient-facing platforms. Left unchecked, these biases risk marginalizing minority populations and reinforcing societal stereotypes.</p>\n </section>\n \n <section>\n \n <h3> Level of Evidence</h3>\n \n <p>4.</p>\n </section>\n </div>","PeriodicalId":48529,"journal":{"name":"Laryngoscope Investigative Otolaryngology","volume":"10 3","pages":""},"PeriodicalIF":1.7000,"publicationDate":"2025-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/lio2.70152","citationCount":"0","resultStr":"{\"title\":\"Representation of Demographics in Otolaryngology by Artificial Intelligence Text-to-Image Platforms\",\"authors\":\"Ariana L. Shaari, Anthony M. Saad, Aman M. Patel, Andrey Filimonov\",\"doi\":\"10.1002/lio2.70152\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n \\n \\n <section>\\n \\n <h3> Objective</h3>\\n \\n <p>Artificial intelligence (AI) text-to-image generators have a propensity to reflect stereotypes. This study investigates the perception of race and gender of AI-generated portraits of otolaryngologists, evaluating their accuracy against workforce demographics and whether they amplify existing social biases.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Methods</h3>\\n \\n <p>Three text-to-image platforms (DALL-E3, Runway, Midjourney) were prompted to generate portrait photos of otolaryngologists based on 29 categories, including personality traits, fellowship, and academic rank. 580 portrait photos were made per platform. Two reviewers characterized the gender and race of the 1740 portraits. Statistical analysis compared the demographics of AI outputs to existing demographic information.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Results</h3>\\n \\n <p>Of the 1740 AI-generated otolaryngologists generated, 88% of images were labeled as White, 4% Black, 6% Asian, 2% Indeterminate/Other race, 88% male, and 12% female. Across academic rank, the representation of White individuals was 97% (department chairs), 90% (program directors), 93% (professors), and 78% (residents). Male representation ranged from 90% (department chairs), 75% (program directors), 100% (professors), and 87% (residents). Runway produced more images of male (89% vs. 88% vs. 85%, <i>p</i> = 0.043) and White (92% vs. 88% vs. 80%, <i>p</i> < 0.001) otolaryngologists than DALL-E3 and Midjourney, respectively.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Conclusion</h3>\\n \\n <p>Text-to-image platforms demonstrated racial and gender biases, with notable differences compared to actual demographics. These platforms often underrepresented females and racial minority groups and overrepresented White males. These disparities underscore the need for the awareness of biases in AI, especially as these tools become more integrated into patient-facing platforms. Left unchecked, these biases risk marginalizing minority populations and reinforcing societal stereotypes.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Level of Evidence</h3>\\n \\n <p>4.</p>\\n </section>\\n </div>\",\"PeriodicalId\":48529,\"journal\":{\"name\":\"Laryngoscope Investigative Otolaryngology\",\"volume\":\"10 3\",\"pages\":\"\"},\"PeriodicalIF\":1.7000,\"publicationDate\":\"2025-05-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1002/lio2.70152\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Laryngoscope Investigative Otolaryngology\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1002/lio2.70152\",\"RegionNum\":4,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"OTORHINOLARYNGOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Laryngoscope Investigative Otolaryngology","FirstCategoryId":"3","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/lio2.70152","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"OTORHINOLARYNGOLOGY","Score":null,"Total":0}
Representation of Demographics in Otolaryngology by Artificial Intelligence Text-to-Image Platforms
Objective
Artificial intelligence (AI) text-to-image generators have a propensity to reflect stereotypes. This study investigates the perception of race and gender of AI-generated portraits of otolaryngologists, evaluating their accuracy against workforce demographics and whether they amplify existing social biases.
Methods
Three text-to-image platforms (DALL-E3, Runway, Midjourney) were prompted to generate portrait photos of otolaryngologists based on 29 categories, including personality traits, fellowship, and academic rank. 580 portrait photos were made per platform. Two reviewers characterized the gender and race of the 1740 portraits. Statistical analysis compared the demographics of AI outputs to existing demographic information.
Results
Of the 1740 AI-generated otolaryngologists generated, 88% of images were labeled as White, 4% Black, 6% Asian, 2% Indeterminate/Other race, 88% male, and 12% female. Across academic rank, the representation of White individuals was 97% (department chairs), 90% (program directors), 93% (professors), and 78% (residents). Male representation ranged from 90% (department chairs), 75% (program directors), 100% (professors), and 87% (residents). Runway produced more images of male (89% vs. 88% vs. 85%, p = 0.043) and White (92% vs. 88% vs. 80%, p < 0.001) otolaryngologists than DALL-E3 and Midjourney, respectively.
Conclusion
Text-to-image platforms demonstrated racial and gender biases, with notable differences compared to actual demographics. These platforms often underrepresented females and racial minority groups and overrepresented White males. These disparities underscore the need for the awareness of biases in AI, especially as these tools become more integrated into patient-facing platforms. Left unchecked, these biases risk marginalizing minority populations and reinforcing societal stereotypes.