Progressive Discriminative Feature Learning for Visible-Infrared Person Re-Identification

Feng Zhou, Zhuxuan Cheng, Haitao Yang, Yifeng Song, Shengpeng Fu
{"title":"Progressive Discriminative Feature Learning for Visible-Infrared Person Re-Identification","authors":"Feng Zhou, Zhuxuan Cheng, Haitao Yang, Yifeng Song, Shengpeng Fu","doi":"10.3390/electronics13142825","DOIUrl":null,"url":null,"abstract":"The visible-infrared person re-identification (VI-ReID) task aims to retrieve the same pedestrian between visible and infrared images. VI-ReID is a challenging task due to the huge modality discrepancy and complex intra-modality variations. Existing works mainly complete the modality alignment at one stage. However, aligning modalities at different stages has positive effects on the intra-class and inter-class distances of cross-modality features, which are often ignored. Moreover, discriminative features with identity information may be corrupted in the processing of modality alignment, further degrading the performance of person re-identification. In this paper, we propose a progressive discriminative feature learning (PDFL) network that adopts different alignment strategies at different stages to alleviate the discrepancy and learn discriminative features progressively. Specifically, we first design an adaptive cross fusion module (ACFM) to learn the identity-relevant features via modality alignment with channel-level attention. For well preserving identity information, we propose a dual-attention-guided instance normalization module (DINM), which can well guide instance normalization to align two modalities into a unified feature space through channel and spatial information embedding. Finally, we generate multiple part features of a person to mine subtle differences. Multi-loss optimization is imposed during the training process for more effective learning supervision. Extensive experiments on the public datasets of SYSU-MM01 and RegDB validate that our proposed method performs favorably against most state-of-the-art methods.","PeriodicalId":504598,"journal":{"name":"Electronics","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Electronics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3390/electronics13142825","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The visible-infrared person re-identification (VI-ReID) task aims to retrieve the same pedestrian between visible and infrared images. VI-ReID is a challenging task due to the huge modality discrepancy and complex intra-modality variations. Existing works mainly complete the modality alignment at one stage. However, aligning modalities at different stages has positive effects on the intra-class and inter-class distances of cross-modality features, which are often ignored. Moreover, discriminative features with identity information may be corrupted in the processing of modality alignment, further degrading the performance of person re-identification. In this paper, we propose a progressive discriminative feature learning (PDFL) network that adopts different alignment strategies at different stages to alleviate the discrepancy and learn discriminative features progressively. Specifically, we first design an adaptive cross fusion module (ACFM) to learn the identity-relevant features via modality alignment with channel-level attention. For well preserving identity information, we propose a dual-attention-guided instance normalization module (DINM), which can well guide instance normalization to align two modalities into a unified feature space through channel and spatial information embedding. Finally, we generate multiple part features of a person to mine subtle differences. Multi-loss optimization is imposed during the training process for more effective learning supervision. Extensive experiments on the public datasets of SYSU-MM01 and RegDB validate that our proposed method performs favorably against most state-of-the-art methods.
用于可见光-红外线人员再识别的渐进式判别特征学习
可见光-红外人员再识别(VI-ReID)任务旨在检索可见光和红外图像中的相同行人。由于存在巨大的模态差异和复杂的模态内变化,VI-ReID 是一项具有挑战性的任务。现有的工作主要是在一个阶段完成模态对齐。然而,在不同阶段进行模态配准对跨模态特征的类内和类间距离有积极影响,而这些影响往往被忽视。此外,带有身份信息的判别特征可能会在模态对齐处理过程中被破坏,从而进一步降低人员再识别的性能。在本文中,我们提出了一种渐进式判别特征学习(PDFL)网络,在不同阶段采用不同的配准策略来缓解差异,并逐步学习判别特征。具体来说,我们首先设计了一个自适应交叉融合模块(ACFM),通过信道级关注的模态配准来学习与身份相关的特征。为了很好地保留身份信息,我们提出了双注意引导的实例归一化模块(DINM),它能很好地引导实例归一化,通过通道和空间信息嵌入将两种模态对齐到统一的特征空间。最后,我们生成一个人的多个部分特征,以挖掘细微差别。在训练过程中,我们进行了多重损失优化,以实现更有效的学习监督。在 SYSU-MM01 和 RegDB 公共数据集上进行的广泛实验验证了我们提出的方法在与大多数最先进方法的比较中表现出色。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信