Fei Zhang;Tianfei Zhou;Jiangchao Yao;Ya Zhang;Ivor W. Tsang;Yanfeng Wang
{"title":"在对齐之前解耦:视觉解耦增强了提示调谐","authors":"Fei Zhang;Tianfei Zhou;Jiangchao Yao;Ya Zhang;Ivor W. Tsang;Yanfeng Wang","doi":"10.1109/TPAMI.2025.3594894","DOIUrl":null,"url":null,"abstract":"<italic>P</i>rompt tuning (PT), as an emerging resource-efficient fine-tuning paradigm, has showcased remarkable effectiveness in improving the task-specific transferability of <italic>vision-language models</i>. This paper delves into a previously overlooked <italic>information asymmetry</i> issue in PT, where the visual modality mostly conveys more context than the object-oriented textual modality. Correspondingly, coarsely aligning these two modalities could result in the <italic>biased attention</i>, driving the model to merely focus on the context area. To address this, we propose DAPT, an effective PT framework based on an intuitive <italic>decouple-before-align</i> concept. First, we propose to explicitly decouple the visual modality into the foreground and background representation via exploiting coarse-and-fine visual segmenting cues, and then both of these decoupled patterns are aligned with the original foreground texts and the hand-crafted background classes, thereby symmetrically strengthening the modal alignment. To further enhance the visual concentration, we propose a visual pull-push regularization tailored for the foreground-background patterns, directing the original visual representation towards unbiased attention on the <italic>region-of-interest</i> object. We demonstrate the power of architecture-free DAPT through <italic>few-shot learning</i>, <italic>base-to-novel generalization</i>, and <italic>data-efficient learning</i>, all of which yield superior performance across prevailing benchmarks.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 11","pages":"10619-10632"},"PeriodicalIF":18.6000,"publicationDate":"2025-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Decouple Before Align: Visual Disentanglement Enhances Prompt Tuning\",\"authors\":\"Fei Zhang;Tianfei Zhou;Jiangchao Yao;Ya Zhang;Ivor W. Tsang;Yanfeng Wang\",\"doi\":\"10.1109/TPAMI.2025.3594894\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<italic>P</i>rompt tuning (PT), as an emerging resource-efficient fine-tuning paradigm, has showcased remarkable effectiveness in improving the task-specific transferability of <italic>vision-language models</i>. This paper delves into a previously overlooked <italic>information asymmetry</i> issue in PT, where the visual modality mostly conveys more context than the object-oriented textual modality. Correspondingly, coarsely aligning these two modalities could result in the <italic>biased attention</i>, driving the model to merely focus on the context area. To address this, we propose DAPT, an effective PT framework based on an intuitive <italic>decouple-before-align</i> concept. First, we propose to explicitly decouple the visual modality into the foreground and background representation via exploiting coarse-and-fine visual segmenting cues, and then both of these decoupled patterns are aligned with the original foreground texts and the hand-crafted background classes, thereby symmetrically strengthening the modal alignment. To further enhance the visual concentration, we propose a visual pull-push regularization tailored for the foreground-background patterns, directing the original visual representation towards unbiased attention on the <italic>region-of-interest</i> object. We demonstrate the power of architecture-free DAPT through <italic>few-shot learning</i>, <italic>base-to-novel generalization</i>, and <italic>data-efficient learning</i>, all of which yield superior performance across prevailing benchmarks.\",\"PeriodicalId\":94034,\"journal\":{\"name\":\"IEEE transactions on pattern analysis and machine intelligence\",\"volume\":\"47 11\",\"pages\":\"10619-10632\"},\"PeriodicalIF\":18.6000,\"publicationDate\":\"2025-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE transactions on pattern analysis and machine intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/11106768/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on pattern analysis and machine intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/11106768/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Decouple Before Align: Visual Disentanglement Enhances Prompt Tuning
Prompt tuning (PT), as an emerging resource-efficient fine-tuning paradigm, has showcased remarkable effectiveness in improving the task-specific transferability of vision-language models. This paper delves into a previously overlooked information asymmetry issue in PT, where the visual modality mostly conveys more context than the object-oriented textual modality. Correspondingly, coarsely aligning these two modalities could result in the biased attention, driving the model to merely focus on the context area. To address this, we propose DAPT, an effective PT framework based on an intuitive decouple-before-align concept. First, we propose to explicitly decouple the visual modality into the foreground and background representation via exploiting coarse-and-fine visual segmenting cues, and then both of these decoupled patterns are aligned with the original foreground texts and the hand-crafted background classes, thereby symmetrically strengthening the modal alignment. To further enhance the visual concentration, we propose a visual pull-push regularization tailored for the foreground-background patterns, directing the original visual representation towards unbiased attention on the region-of-interest object. We demonstrate the power of architecture-free DAPT through few-shot learning, base-to-novel generalization, and data-efficient learning, all of which yield superior performance across prevailing benchmarks.