LSTIF:Long-short Temporal Information Fusion Architecture for Video-based Person Re-identification

Xingzhe Sun, Shanna Zhuang, Zhengyou Wang
{"title":"LSTIF:Long-short Temporal Information Fusion Architecture for Video-based Person Re-identification","authors":"Xingzhe Sun, Shanna Zhuang, Zhengyou Wang","doi":"10.1109/ICCEAI52939.2021.00027","DOIUrl":null,"url":null,"abstract":"Person re-identification is a major application of computer vision in reality. Since the data obtained by monitoring in real life is often in video format, and the walking poses of pedestrians are different, in addition to the appearance of pedestrians, how to obtain the motion features of pedestrians, is extremely important for video-based person re-identification. Therefore, for the temporal information of the video, we propose a Long-short Temporal Information Fusion (LSTIF) network. We aggregate temporal information from two perspectives, short-term features containing detailed information and long-term features containing global information. Simultaneously, in order to reduce the amount of calculation, this network also uses non-local blocks, and extend the outpu feature map to the same size as the input, which is convenient for calculation. This paper verifies the effectiveness of our method on two commonly used datasets iLIDS-VID and DukeMTMC-VideoReID.","PeriodicalId":331409,"journal":{"name":"2021 International Conference on Computer Engineering and Artificial Intelligence (ICCEAI)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2021-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Computer Engineering and Artificial Intelligence (ICCEAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCEAI52939.2021.00027","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Person re-identification is a major application of computer vision in reality. Since the data obtained by monitoring in real life is often in video format, and the walking poses of pedestrians are different, in addition to the appearance of pedestrians, how to obtain the motion features of pedestrians, is extremely important for video-based person re-identification. Therefore, for the temporal information of the video, we propose a Long-short Temporal Information Fusion (LSTIF) network. We aggregate temporal information from two perspectives, short-term features containing detailed information and long-term features containing global information. Simultaneously, in order to reduce the amount of calculation, this network also uses non-local blocks, and extend the outpu feature map to the same size as the input, which is convenient for calculation. This paper verifies the effectiveness of our method on two commonly used datasets iLIDS-VID and DukeMTMC-VideoReID.
基于视频的人物再识别的长-短时间信息融合体系结构
人的再识别是计算机视觉在现实中的一个重要应用。由于现实生活中监控获得的数据往往是视频格式的,而行人的行走姿势又各不相同,所以除了行人的外观外,如何获取行人的运动特征,对于基于视频的人再识别来说是极其重要的。因此,针对视频的时间信息,我们提出了一种长-短时间信息融合(LSTIF)网络。我们从两个角度聚合时间信息,包含详细信息的短期特征和包含全局信息的长期特征。同时,为了减少计算量,该网络还使用了非局部块,并将输出特征映射扩展到与输入相同的大小,方便计算。本文在两个常用数据集iLIDS-VID和DukeMTMC-VideoReID上验证了该方法的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信