{"title":"AGFF: Attention-Gated Feature Fusion for Multi-Pose Virtual Try-On","authors":"Chenghu Du;Peiliang Zhang;Junyin Wang;Shengwu Xiong","doi":"10.1109/TCE.2025.3535237","DOIUrl":null,"url":null,"abstract":"Image-based multi-pose virtual try-on tasks aim to synthesize a person wearing a garment in a desired posture. Current methods face three main challenges: i) natural warping of clothing images, ii) preserving identity information after pose transfer, and iii) reducing model complexity. While recent methods have made improvements, they struggle to trade off complexity and performance, limiting model generalization. To address these issues, we propose AGFF, a new image-based multi-pose virtual try-on network with <bold>A</b>ttention-<bold>G</b>ated <bold>F</b>eature <bold>F</b>usion (AGFF), which efficiently tries on garments in arbitrary poses with low complexity. First, we introduce a bi-directional feature-matching approach with feature warping to capture geometric matching information between the garment and human posture for complex posture alignment. Second, we propose an attention-gated feature fusion approach to preserve more identity information by suppressing irrelevant person features and enhancing salient ones. Additionally, our model integrates seamlessly into small-scale encoder-decoder architectures, further reducing complexity. Extensive experiments on popular benchmarks show that our method outperforms state-of-the-art approaches both qualitatively and quantitatively.","PeriodicalId":13208,"journal":{"name":"IEEE Transactions on Consumer Electronics","volume":"71 1","pages":"819-827"},"PeriodicalIF":4.3000,"publicationDate":"2025-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Consumer Electronics","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10855617/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Image-based multi-pose virtual try-on tasks aim to synthesize a person wearing a garment in a desired posture. Current methods face three main challenges: i) natural warping of clothing images, ii) preserving identity information after pose transfer, and iii) reducing model complexity. While recent methods have made improvements, they struggle to trade off complexity and performance, limiting model generalization. To address these issues, we propose AGFF, a new image-based multi-pose virtual try-on network with Attention-Gated Feature Fusion (AGFF), which efficiently tries on garments in arbitrary poses with low complexity. First, we introduce a bi-directional feature-matching approach with feature warping to capture geometric matching information between the garment and human posture for complex posture alignment. Second, we propose an attention-gated feature fusion approach to preserve more identity information by suppressing irrelevant person features and enhancing salient ones. Additionally, our model integrates seamlessly into small-scale encoder-decoder architectures, further reducing complexity. Extensive experiments on popular benchmarks show that our method outperforms state-of-the-art approaches both qualitatively and quantitatively.
期刊介绍:
The main focus for the IEEE Transactions on Consumer Electronics is the engineering and research aspects of the theory, design, construction, manufacture or end use of mass market electronics, systems, software and services for consumers.