rgb -红外人员再识别的跨模态信道混频与模态去相关

Boyu Hua;Junyin Zhang;Ziqiang Li;Yongxin Ge
{"title":"rgb -红外人员再识别的跨模态信道混频与模态去相关","authors":"Boyu Hua;Junyin Zhang;Ziqiang Li;Yongxin Ge","doi":"10.1109/TBIOM.2023.3287275","DOIUrl":null,"url":null,"abstract":"This paper focuses on RGB-infrared person re-identification, which is challenged by a large modality gap between RGB and infrared images. Most existing methods attempt to learn discriminative modality-invariant features. These methods make use of identity annotations while they do not sufficiently exploit intra-modality and cross-modality sample relations using modality annotations. In this paper, we propose a Cross-modality channel Mixup and Modality Decorrelation method (CMMD) that explores sample relations at both image and feature levels. This method is designed to reduce redundant modality-specific information of the representations and highlight modality-shared information. Specifically, we first design a cross-modality channel mixup (CCM) augmentation at the image level, which combines a random RGB channel and an infrared image to generate a new one by mixup, while keeping identity information unchanged. This augmentation can be integrated into other methods easily without introducing extra parameters or models. In addition, modality decorrelation quintuplet loss (MDQL) is further presented to mine hard samples in a batch, that is, positive/negative intra/cross-modality samples, to learn modality-invariant representations in the shared latent space at the feature level. This loss suggests that the closest negative sample and the farthest positive sample should have an equal probability of appearing in both modalities. Comprehensive experimental results on two challenging datasets, i.e., SYSY-MM01 and RegDB, demonstrate competitive performance of our method with state-of-the-art ones.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"5 4","pages":"512-523"},"PeriodicalIF":0.0000,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Cross-Modality Channel Mixup and Modality Decorrelation for RGB-Infrared Person Re-Identification\",\"authors\":\"Boyu Hua;Junyin Zhang;Ziqiang Li;Yongxin Ge\",\"doi\":\"10.1109/TBIOM.2023.3287275\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper focuses on RGB-infrared person re-identification, which is challenged by a large modality gap between RGB and infrared images. Most existing methods attempt to learn discriminative modality-invariant features. These methods make use of identity annotations while they do not sufficiently exploit intra-modality and cross-modality sample relations using modality annotations. In this paper, we propose a Cross-modality channel Mixup and Modality Decorrelation method (CMMD) that explores sample relations at both image and feature levels. This method is designed to reduce redundant modality-specific information of the representations and highlight modality-shared information. Specifically, we first design a cross-modality channel mixup (CCM) augmentation at the image level, which combines a random RGB channel and an infrared image to generate a new one by mixup, while keeping identity information unchanged. This augmentation can be integrated into other methods easily without introducing extra parameters or models. In addition, modality decorrelation quintuplet loss (MDQL) is further presented to mine hard samples in a batch, that is, positive/negative intra/cross-modality samples, to learn modality-invariant representations in the shared latent space at the feature level. This loss suggests that the closest negative sample and the farthest positive sample should have an equal probability of appearing in both modalities. Comprehensive experimental results on two challenging datasets, i.e., SYSY-MM01 and RegDB, demonstrate competitive performance of our method with state-of-the-art ones.\",\"PeriodicalId\":73307,\"journal\":{\"name\":\"IEEE transactions on biometrics, behavior, and identity science\",\"volume\":\"5 4\",\"pages\":\"512-523\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-06-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE transactions on biometrics, behavior, and identity science\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10163902/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on biometrics, behavior, and identity science","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10163902/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

针对RGB图像与红外图像存在较大模态差异的问题,本文重点研究了RGB-红外图像对人的再识别。大多数现有的方法试图学习判别模态不变特征。这些方法利用了身份注释,但没有充分利用模态注释的模态内和跨模态样本关系。在本文中,我们提出了一种跨模态通道混合和模态去相关方法(CMMD),该方法在图像和特征级别上探索样本关系。该方法旨在减少表示中冗余的特定于模态的信息,并突出显示模态共享信息。具体而言,我们首先在图像级设计了一种交叉模态通道混合(CCM)增强,该增强将随机RGB通道与红外图像结合在一起,在保持身份信息不变的情况下通过混合生成新的通道。这种增强可以很容易地集成到其他方法中,而无需引入额外的参数或模型。此外,进一步提出了模态去相关五元损失(MDQL),用于批量挖掘硬样本,即正/负模态内/跨模态样本,以在特征层面学习共享潜在空间中的模态不变表示。这种损失表明,最近的阴性样本和最远的阳性样本在两种模式中出现的概率应该相等。在SYSY-MM01和RegDB两个具有挑战性的数据集上的综合实验结果表明,我们的方法与最先进的方法相比具有竞争力。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Cross-Modality Channel Mixup and Modality Decorrelation for RGB-Infrared Person Re-Identification
This paper focuses on RGB-infrared person re-identification, which is challenged by a large modality gap between RGB and infrared images. Most existing methods attempt to learn discriminative modality-invariant features. These methods make use of identity annotations while they do not sufficiently exploit intra-modality and cross-modality sample relations using modality annotations. In this paper, we propose a Cross-modality channel Mixup and Modality Decorrelation method (CMMD) that explores sample relations at both image and feature levels. This method is designed to reduce redundant modality-specific information of the representations and highlight modality-shared information. Specifically, we first design a cross-modality channel mixup (CCM) augmentation at the image level, which combines a random RGB channel and an infrared image to generate a new one by mixup, while keeping identity information unchanged. This augmentation can be integrated into other methods easily without introducing extra parameters or models. In addition, modality decorrelation quintuplet loss (MDQL) is further presented to mine hard samples in a batch, that is, positive/negative intra/cross-modality samples, to learn modality-invariant representations in the shared latent space at the feature level. This loss suggests that the closest negative sample and the farthest positive sample should have an equal probability of appearing in both modalities. Comprehensive experimental results on two challenging datasets, i.e., SYSY-MM01 and RegDB, demonstrate competitive performance of our method with state-of-the-art ones.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
10.90
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信