{"title":"Pedestrian fusion tracking method based on multimodal information com-plementation","authors":"Zhang Xue, Li Yi, Zuo Jie, Liu Shiqian","doi":"10.1109/ICCEA53728.2021.00028","DOIUrl":null,"url":null,"abstract":"A pedestrian fusion tracking method with complementary multimodal information is proposed, and a fusion decision tracking model with detection followed by fusion and then tracking is established. The detection module uses a modified CenterNet network with a richer feature information backbone network and a lightweight prediction module, and the scene data is collected to train multiple detectors for multiple modalities. A decision process based on the confidence of detection results and feature similarity is proposed to achieve the fusion of multimodal detection results, and the fused results are fed to the tracker to achieve continuous pedestrian tracking. The results show that the proposed fusion tracking model can complement each other’s multi-modal information and provide better and more robust tracking results than the single-modal tracker for continuous tracking in multiple scenes.","PeriodicalId":325790,"journal":{"name":"2021 International Conference on Computer Engineering and Application (ICCEA)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Computer Engineering and Application (ICCEA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCEA53728.2021.00028","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
A pedestrian fusion tracking method with complementary multimodal information is proposed, and a fusion decision tracking model with detection followed by fusion and then tracking is established. The detection module uses a modified CenterNet network with a richer feature information backbone network and a lightweight prediction module, and the scene data is collected to train multiple detectors for multiple modalities. A decision process based on the confidence of detection results and feature similarity is proposed to achieve the fusion of multimodal detection results, and the fused results are fed to the tracker to achieve continuous pedestrian tracking. The results show that the proposed fusion tracking model can complement each other’s multi-modal information and provide better and more robust tracking results than the single-modal tracker for continuous tracking in multiple scenes.