{"title":"基于视频的人物再识别多尺度特征融合网络","authors":"Penggao Liu, M. Ai, Guozhi Shan","doi":"10.1109/ICETCI53161.2021.9563438","DOIUrl":null,"url":null,"abstract":"In recent years, person re-identification technology has been greatly developed. Image-based person re-identification algorithms have achieved excellent performance on open source datasets. In contrast, the development of video-based person re-identification technology is relatively backward. At present, the main research work of video-based person re-identification algorithms is focused on the processing of temporal information in the picture sequence. Complex appearance features are not effective when performing temporal fusion, so the frame-level features used are almost based on global features. This paper proposes a video person re-identification model based on multi-scale feature fusion. The multi-scale feature fusion of the model is embodied in the design of the frame-level feature extraction module. This module extracts the frame-level features of different scales, and then catenates them together into vectors, which not only improves the feature discrimination degree, but also makes the catenated frame-level features carry out effective temporal fusion, and the test results on the Mars dataset have reached a competitive level. At the same time, a series of comparative experiments were carried out on the model parameters to achieve further optimization of performance.","PeriodicalId":170858,"journal":{"name":"2021 IEEE International Conference on Electronic Technology, Communication and Information (ICETCI)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multi-Scale Feature Fusion Network for Video-Based Person Re-Identification\",\"authors\":\"Penggao Liu, M. Ai, Guozhi Shan\",\"doi\":\"10.1109/ICETCI53161.2021.9563438\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In recent years, person re-identification technology has been greatly developed. Image-based person re-identification algorithms have achieved excellent performance on open source datasets. In contrast, the development of video-based person re-identification technology is relatively backward. At present, the main research work of video-based person re-identification algorithms is focused on the processing of temporal information in the picture sequence. Complex appearance features are not effective when performing temporal fusion, so the frame-level features used are almost based on global features. This paper proposes a video person re-identification model based on multi-scale feature fusion. The multi-scale feature fusion of the model is embodied in the design of the frame-level feature extraction module. This module extracts the frame-level features of different scales, and then catenates them together into vectors, which not only improves the feature discrimination degree, but also makes the catenated frame-level features carry out effective temporal fusion, and the test results on the Mars dataset have reached a competitive level. At the same time, a series of comparative experiments were carried out on the model parameters to achieve further optimization of performance.\",\"PeriodicalId\":170858,\"journal\":{\"name\":\"2021 IEEE International Conference on Electronic Technology, Communication and Information (ICETCI)\",\"volume\":\"48 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-08-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE International Conference on Electronic Technology, Communication and Information (ICETCI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICETCI53161.2021.9563438\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Conference on Electronic Technology, Communication and Information (ICETCI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICETCI53161.2021.9563438","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Multi-Scale Feature Fusion Network for Video-Based Person Re-Identification
In recent years, person re-identification technology has been greatly developed. Image-based person re-identification algorithms have achieved excellent performance on open source datasets. In contrast, the development of video-based person re-identification technology is relatively backward. At present, the main research work of video-based person re-identification algorithms is focused on the processing of temporal information in the picture sequence. Complex appearance features are not effective when performing temporal fusion, so the frame-level features used are almost based on global features. This paper proposes a video person re-identification model based on multi-scale feature fusion. The multi-scale feature fusion of the model is embodied in the design of the frame-level feature extraction module. This module extracts the frame-level features of different scales, and then catenates them together into vectors, which not only improves the feature discrimination degree, but also makes the catenated frame-level features carry out effective temporal fusion, and the test results on the Mars dataset have reached a competitive level. At the same time, a series of comparative experiments were carried out on the model parameters to achieve further optimization of performance.