Rethinking Temporal Object Detection from Robotic Perspectives

Xingyu Chen, Zhengxing Wu, Junzhi Yu, Li Wen
{"title":"Rethinking Temporal Object Detection from Robotic Perspectives","authors":"Xingyu Chen, Zhengxing Wu, Junzhi Yu, Li Wen","doi":"10.1201/9781003144281-5","DOIUrl":null,"url":null,"abstract":"Video object detection (VID) has been vigorously studied for years but almost all literature adopts a static accuracy-based evaluation, i.e., average precision (AP). From a robotic perspective, the importance of recall continuity and localization stability is equal to that of accuracy, but the AP is insufficient to reflect detectors' performance across time. In this paper, non-reference assessments are proposed for continuity and stability based on object tracklets. These temporal evaluations can serve as supplements to static AP. Further, we develop an online tracklet refinement for improving detectors' temporal performance through short tracklet suppression, fragment filling, and temporal location fusion. \nIn addition, we propose a small-overlap suppression to extend VID methods to single object tracking (SOT) task so that a flexible SOT-by-detection framework is then formed. \nExtensive experiments are conducted on ImageNet VID dataset and real-world robotic tasks, where the superiority of our proposed approaches are validated and verified. Codes will be publicly available.","PeriodicalId":413618,"journal":{"name":"Visual Perception and Control of Underwater Robots","volume":"50 3 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Visual Perception and Control of Underwater Robots","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1201/9781003144281-5","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

Video object detection (VID) has been vigorously studied for years but almost all literature adopts a static accuracy-based evaluation, i.e., average precision (AP). From a robotic perspective, the importance of recall continuity and localization stability is equal to that of accuracy, but the AP is insufficient to reflect detectors' performance across time. In this paper, non-reference assessments are proposed for continuity and stability based on object tracklets. These temporal evaluations can serve as supplements to static AP. Further, we develop an online tracklet refinement for improving detectors' temporal performance through short tracklet suppression, fragment filling, and temporal location fusion. In addition, we propose a small-overlap suppression to extend VID methods to single object tracking (SOT) task so that a flexible SOT-by-detection framework is then formed. Extensive experiments are conducted on ImageNet VID dataset and real-world robotic tasks, where the superiority of our proposed approaches are validated and verified. Codes will be publicly available.
从机器人角度重新思考时间目标检测
视频目标检测(VID)已被大力研究多年,但几乎所有文献都采用基于静态精度的评价,即平均精度(AP)。从机器人的角度来看,召回连续性和定位稳定性的重要性与准确性相等,但AP不足以反映探测器随时间的性能。本文提出了一种基于目标轨迹的连续性和稳定性非参考评价方法。这些时间评估可以作为静态AP的补充。此外,我们开发了一种在线轨道优化方法,通过短轨道抑制、碎片填充和时间位置融合来提高检测器的时间性能。此外,我们提出了一种小重叠抑制方法,将VID方法扩展到单目标跟踪(SOT)任务,从而形成一个灵活的SOT-by-detection框架。在ImageNet VID数据集和现实世界的机器人任务上进行了大量的实验,验证了我们提出的方法的优越性。代码将公开提供。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信