Yabin Zhu , Qianwu Wang , Chenglong Li , Jin Tang , Chengjie Gu , Zhixiang Huang
{"title":"Visible–thermal multiple object tracking: Large-scale video dataset and progressive fusion approach","authors":"Yabin Zhu , Qianwu Wang , Chenglong Li , Jin Tang , Chengjie Gu , Zhixiang Huang","doi":"10.1016/j.patcog.2024.111330","DOIUrl":null,"url":null,"abstract":"<div><div>The complementary benefits from visible and thermal infrared data are extensively utilized in various computer vision tasks, such as visual tracking and object detection, but rarely explored in Multiple Object Tracking (MOT). This paper contributes a large-scale Visible–Thermal video benchmark for MOT, named VT-MOT, which presents several key advantages. First, it comprises 582 video sequence pairs with 401,000 frame pairs collected from diverse sources, including surveillance, drone, and handheld platforms. Second, VT-MOT has dense and high-quality annotations, with 3.99 million annotation boxes verified by professionals. To provide a strong baseline, we design a simple yet effective tracking framework, which effectively fuses temporal information and complementary information of two modalities in a progressive manner, for robust visible–thermal MOT. Comprehensive experiments validate the proposed method’s superiority over existing state-of-the-art methods, while potential future research directions for visible–thermal MOT are outlined. The project is released in <span><span>https://github.com/wqw123wqw/PFTrack</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"161 ","pages":"Article 111330"},"PeriodicalIF":7.5000,"publicationDate":"2024-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Pattern Recognition","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0031320324010811","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
The complementary benefits from visible and thermal infrared data are extensively utilized in various computer vision tasks, such as visual tracking and object detection, but rarely explored in Multiple Object Tracking (MOT). This paper contributes a large-scale Visible–Thermal video benchmark for MOT, named VT-MOT, which presents several key advantages. First, it comprises 582 video sequence pairs with 401,000 frame pairs collected from diverse sources, including surveillance, drone, and handheld platforms. Second, VT-MOT has dense and high-quality annotations, with 3.99 million annotation boxes verified by professionals. To provide a strong baseline, we design a simple yet effective tracking framework, which effectively fuses temporal information and complementary information of two modalities in a progressive manner, for robust visible–thermal MOT. Comprehensive experiments validate the proposed method’s superiority over existing state-of-the-art methods, while potential future research directions for visible–thermal MOT are outlined. The project is released in https://github.com/wqw123wqw/PFTrack.
期刊介绍:
The field of Pattern Recognition is both mature and rapidly evolving, playing a crucial role in various related fields such as computer vision, image processing, text analysis, and neural networks. It closely intersects with machine learning and is being applied in emerging areas like biometrics, bioinformatics, multimedia data analysis, and data science. The journal Pattern Recognition, established half a century ago during the early days of computer science, has since grown significantly in scope and influence.