Feature Fusion with Non-local for Pedestrian Attribute Recognition

Jiping Lv, Zhenghua Xiong, Rongfang Zou, Zhangying Wen, Hongli Lin
{"title":"Feature Fusion with Non-local for Pedestrian Attribute Recognition","authors":"Jiping Lv, Zhenghua Xiong, Rongfang Zou, Zhangying Wen, Hongli Lin","doi":"10.1145/3523286.3524581","DOIUrl":null,"url":null,"abstract":"In the Pedestrian Attribute Recognition (PAR) research topic, how to extract comprehensive features to represent pedestrian attributes is still an open problem due to its multi-label nature. In order to tackle this problem, we proposed a new end-to-end PAR network based on Feature Fusion with a Non-local operation(FFNL), named FFNL-net. Compared with existing feature fusion approaches that only mechanically pays attention to feature maps from multiple levels, it also exploits the strongest semantic but still higher resolution feature maps. Firstly, the pedestrian image is extracted from a backbone network. Then, three feature maps from different levels and scales are obtained, and the Non-local operation on the above multiple feature maps are applied respectively to further extract spatial information. Finally, a novel feature fusion strategy is performed on them to fetch a fused feature map. Our experimental results on existing popular pedestrian attribute datasets of PETA, PA-100K, and RAP prove that our proposed approach achieves state-of-the-art results.","PeriodicalId":268165,"journal":{"name":"2022 2nd International Conference on Bioinformatics and Intelligent Computing","volume":"27 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 2nd International Conference on Bioinformatics and Intelligent Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3523286.3524581","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

In the Pedestrian Attribute Recognition (PAR) research topic, how to extract comprehensive features to represent pedestrian attributes is still an open problem due to its multi-label nature. In order to tackle this problem, we proposed a new end-to-end PAR network based on Feature Fusion with a Non-local operation(FFNL), named FFNL-net. Compared with existing feature fusion approaches that only mechanically pays attention to feature maps from multiple levels, it also exploits the strongest semantic but still higher resolution feature maps. Firstly, the pedestrian image is extracted from a backbone network. Then, three feature maps from different levels and scales are obtained, and the Non-local operation on the above multiple feature maps are applied respectively to further extract spatial information. Finally, a novel feature fusion strategy is performed on them to fetch a fused feature map. Our experimental results on existing popular pedestrian attribute datasets of PETA, PA-100K, and RAP prove that our proposed approach achieves state-of-the-art results.
非局部特征融合行人属性识别
在行人属性识别(PAR)研究课题中,由于行人属性的多标签特性,如何提取综合特征来表示行人属性一直是一个有待解决的问题。为了解决这一问题,我们提出了一种基于特征融合和非局部操作(FFNL)的端到端PAR网络,命名为FFNL-net。与现有的特征融合方法只机械地关注多个层次的特征图相比,该方法利用了语义最强但分辨率更高的特征图。首先,从主干网络中提取行人图像;然后,得到三个不同层次和尺度的特征图,分别对上述多个特征图进行非局部运算,进一步提取空间信息。最后,采用一种新的特征融合策略对它们进行融合,得到融合后的特征映射。我们在PETA、PA-100K和RAP的现有流行行人属性数据集上的实验结果证明了我们的方法达到了最先进的结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信