Yuqi Li , Yanli Li , Kai Zhang , Fuyan Zhang , Chuanguang Yang , Zhongliang Guo , Weiping Ding , Tingwen Huang
{"title":"Achieving fair medical image segmentation in foundation models with adversarial visual prompt tuning","authors":"Yuqi Li , Yanli Li , Kai Zhang , Fuyan Zhang , Chuanguang Yang , Zhongliang Guo , Weiping Ding , Tingwen Huang","doi":"10.1016/j.ins.2025.122501","DOIUrl":null,"url":null,"abstract":"<div><div>Recent advances in deep learning have significantly enhanced medical image analysis capabilities. Medical image segmentation, a critical application in this domain, enables precise delineation of anatomical structures and pathological regions, substantially supporting clinical decision-making. However, current segmentation methods primarily optimize for overall performance without considering disparities across demographic groups, raising important fairness concerns. To address this gap, we propose <strong><u>Adv</u></strong>ersarial <strong><u>V</u></strong>isual <strong><u>P</u></strong>rompt <strong><u>T</u></strong>uning (<strong>AdvVPT</strong>), a parameter-efficient approach that enhances fairness in foundation models for medical image segmentation. AdvVPT introduces trainable visual prompts within the image encoder while keeping the backbone frozen, requiring only 0.812M additional parameters. These prompts are optimized through adversarial training to absorb demographic-specific biased information from image embeddings, achieved by maximizing prediction errors for sensitive attributes and increasing embedding distances between visual prompts and image features. Experimental evaluation on the Harvard-FairSeg dataset demonstrates that AdvVPT achieves state-of-the-art fairness performance across multiple demographic attributes. For racial fairness, AdvVPT achieves an ES-Dice score of 0.8996 and an ES-IoU score of 0.8222 on optic cup segmentation, substantially outperforming existing methods. For gender fairness using the SAT backbone, AdvVPT achieves an ES-Dice of 0.9297 and ES-IoU of 0.8614, demonstrating both superior performance and improved balance between male and female subgroups.</div></div>","PeriodicalId":51063,"journal":{"name":"Information Sciences","volume":"720 ","pages":"Article 122501"},"PeriodicalIF":6.8000,"publicationDate":"2025-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Sciences","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0020025525006334","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"0","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Recent advances in deep learning have significantly enhanced medical image analysis capabilities. Medical image segmentation, a critical application in this domain, enables precise delineation of anatomical structures and pathological regions, substantially supporting clinical decision-making. However, current segmentation methods primarily optimize for overall performance without considering disparities across demographic groups, raising important fairness concerns. To address this gap, we propose Adversarial Visual Prompt Tuning (AdvVPT), a parameter-efficient approach that enhances fairness in foundation models for medical image segmentation. AdvVPT introduces trainable visual prompts within the image encoder while keeping the backbone frozen, requiring only 0.812M additional parameters. These prompts are optimized through adversarial training to absorb demographic-specific biased information from image embeddings, achieved by maximizing prediction errors for sensitive attributes and increasing embedding distances between visual prompts and image features. Experimental evaluation on the Harvard-FairSeg dataset demonstrates that AdvVPT achieves state-of-the-art fairness performance across multiple demographic attributes. For racial fairness, AdvVPT achieves an ES-Dice score of 0.8996 and an ES-IoU score of 0.8222 on optic cup segmentation, substantially outperforming existing methods. For gender fairness using the SAT backbone, AdvVPT achieves an ES-Dice of 0.9297 and ES-IoU of 0.8614, demonstrating both superior performance and improved balance between male and female subgroups.
期刊介绍:
Informatics and Computer Science Intelligent Systems Applications is an esteemed international journal that focuses on publishing original and creative research findings in the field of information sciences. We also feature a limited number of timely tutorial and surveying contributions.
Our journal aims to cater to a diverse audience, including researchers, developers, managers, strategic planners, graduate students, and anyone interested in staying up-to-date with cutting-edge research in information science, knowledge engineering, and intelligent systems. While readers are expected to share a common interest in information science, they come from varying backgrounds such as engineering, mathematics, statistics, physics, computer science, cell biology, molecular biology, management science, cognitive science, neurobiology, behavioral sciences, and biochemistry.