FACT: Feature Adaptive Continual-learning Tracker for Multiple Object Tracking

Rongzihan Song, Zhenyu Weng, Huiping Zhuang, Jinchang Ren, Yongming Chen, Zhiping Lin
{"title":"FACT: Feature Adaptive Continual-learning Tracker for Multiple Object Tracking","authors":"Rongzihan Song, Zhenyu Weng, Huiping Zhuang, Jinchang Ren, Yongming Chen, Zhiping Lin","doi":"arxiv-2409.07904","DOIUrl":null,"url":null,"abstract":"Multiple object tracking (MOT) involves identifying multiple targets and\nassigning them corresponding IDs within a video sequence, where occlusions are\noften encountered. Recent methods address occlusions using appearance cues\nthrough online learning techniques to improve adaptivity or offline learning\ntechniques to utilize temporal information from videos. However, most existing\nonline learning-based MOT methods are unable to learn from all past tracking\ninformation to improve adaptivity on long-term occlusions while maintaining\nreal-time tracking speed. On the other hand, temporal information-based offline\nlearning methods maintain a long-term memory to store past tracking\ninformation, but this approach restricts them to use only local past\ninformation during tracking. To address these challenges, we propose a new MOT\nframework called the Feature Adaptive Continual-learning Tracker (FACT), which\nenables real-time tracking and feature learning for targets by utilizing all\npast tracking information. We demonstrate that the framework can be integrated\nwith various state-of-the-art feature-based trackers, thereby improving their\ntracking ability. Specifically, we develop the feature adaptive\ncontinual-learning (FAC) module, a neural network that can be trained online to\nlearn features adaptively using all past tracking information during tracking.\nMoreover, we also introduce a two-stage association module specifically\ndesigned for the proposed continual learning-based tracking. Extensive\nexperiment results demonstrate that the proposed method achieves\nstate-of-the-art online tracking performance on MOT17 and MOT20 benchmarks. The\ncode will be released upon acceptance.","PeriodicalId":501130,"journal":{"name":"arXiv - CS - Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computer Vision and Pattern Recognition","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07904","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Multiple object tracking (MOT) involves identifying multiple targets and assigning them corresponding IDs within a video sequence, where occlusions are often encountered. Recent methods address occlusions using appearance cues through online learning techniques to improve adaptivity or offline learning techniques to utilize temporal information from videos. However, most existing online learning-based MOT methods are unable to learn from all past tracking information to improve adaptivity on long-term occlusions while maintaining real-time tracking speed. On the other hand, temporal information-based offline learning methods maintain a long-term memory to store past tracking information, but this approach restricts them to use only local past information during tracking. To address these challenges, we propose a new MOT framework called the Feature Adaptive Continual-learning Tracker (FACT), which enables real-time tracking and feature learning for targets by utilizing all past tracking information. We demonstrate that the framework can be integrated with various state-of-the-art feature-based trackers, thereby improving their tracking ability. Specifically, we develop the feature adaptive continual-learning (FAC) module, a neural network that can be trained online to learn features adaptively using all past tracking information during tracking. Moreover, we also introduce a two-stage association module specifically designed for the proposed continual learning-based tracking. Extensive experiment results demonstrate that the proposed method achieves state-of-the-art online tracking performance on MOT17 and MOT20 benchmarks. The code will be released upon acceptance.
FACT:用于多目标跟踪的特征自适应持续学习跟踪器
多目标跟踪(MOT)涉及在视频序列中识别多个目标并为其分配相应的 ID,而在视频序列中经常会遇到遮挡物。最近的方法通过在线学习技术来提高适应性,或通过离线学习技术来利用视频中的时间信息,从而利用外观线索来解决遮挡问题。然而,大多数现有的基于在线学习的 MOT 方法都无法从所有过去的跟踪信息中学习,从而在保持实时跟踪速度的同时提高对长期遮挡的适应性。另一方面,基于时间信息的离线学习方法会保留一个长期存储器来存储过去的跟踪信息,但这种方法限制了它们在跟踪过程中只能使用局部的过去信息。为了应对这些挑战,我们提出了一种名为 "特征自适应持续学习跟踪器"(FACT)的新型 MOT 框架,通过利用所有过去的跟踪信息,实现对目标的实时跟踪和特征学习。我们证明,该框架可以与各种最先进的基于特征的跟踪器集成,从而提高它们的跟踪能力。具体来说,我们开发了特征自适应持续学习(FAC)模块,这是一个可在线训练的神经网络,可在跟踪过程中利用所有过去的跟踪信息自适应地学习特征。广泛的实验结果表明,所提出的方法在 MOT17 和 MOT20 基准上实现了最先进的在线跟踪性能。代码将在验收通过后发布。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信