VRGNet: A Robust Visible Region-Guided Network for Occluded Pedestrian Detection

Xin Mao, Chaoqi Yan, Hong Zhang, J. Song, Ding Yuan
{"title":"VRGNet: A Robust Visible Region-Guided Network for Occluded Pedestrian Detection","authors":"Xin Mao, Chaoqi Yan, Hong Zhang, J. Song, Ding Yuan","doi":"10.1145/3581807.3581817","DOIUrl":null,"url":null,"abstract":"Pedestrian detection has made significant progress in both academic and industrial fields. However, there are still some challenging questions with regard to occlusion scene. In this paper, we propose a novel and robust visible region-guided network (VRGNet) to specially improve the occluded pedestrian detection performance. Specifically, we leverage the adapted FPN-based framework to extract multi-scale features, and fuse them together to encode more precision localization and semantic information. In addition, we construct a pedestrian part pool that covers almost all the scale of different occluded body regions. Meanwhile, we propose a new occlusion handling strategy by elaborately integrating the prior knowledge of different visible body regions with visibility prediction into the detection framework to deal with pedestrians with different degree of occlusion. The extensive experiments demonstrate that our VRGNet achieves a leading performance under different evaluation settings on Caltech-USA dataset, especially for occluded pedestrians. In addition, it also achieves a competitive of 48.4%, 9.3%, 6.7% under the Heavy, Partial and Bare settings respectively on CityPersons dataset compared with other state-of-the-art pedestrian detection algorithms, while keeping a better speed-accuracy trade-off.","PeriodicalId":292813,"journal":{"name":"Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition","volume":"15 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3581807.3581817","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Pedestrian detection has made significant progress in both academic and industrial fields. However, there are still some challenging questions with regard to occlusion scene. In this paper, we propose a novel and robust visible region-guided network (VRGNet) to specially improve the occluded pedestrian detection performance. Specifically, we leverage the adapted FPN-based framework to extract multi-scale features, and fuse them together to encode more precision localization and semantic information. In addition, we construct a pedestrian part pool that covers almost all the scale of different occluded body regions. Meanwhile, we propose a new occlusion handling strategy by elaborately integrating the prior knowledge of different visible body regions with visibility prediction into the detection framework to deal with pedestrians with different degree of occlusion. The extensive experiments demonstrate that our VRGNet achieves a leading performance under different evaluation settings on Caltech-USA dataset, especially for occluded pedestrians. In addition, it also achieves a competitive of 48.4%, 9.3%, 6.7% under the Heavy, Partial and Bare settings respectively on CityPersons dataset compared with other state-of-the-art pedestrian detection algorithms, while keeping a better speed-accuracy trade-off.
VRGNet:用于遮挡行人检测的鲁棒可见区域引导网络
行人检测在学术和工业领域都取得了重大进展。然而,关于遮挡场景,仍然存在一些具有挑战性的问题。在本文中,我们提出了一种新颖的鲁棒可见区域引导网络(VRGNet)来提高遮挡行人的检测性能。具体来说,我们利用改进的基于fpn的框架来提取多尺度特征,并将它们融合在一起以编码更精确的定位和语义信息。此外,我们构建了一个行人部分池,几乎涵盖了不同遮挡体区域的所有尺度。同时,我们提出了一种新的遮挡处理策略,将不同可见身体区域的先验知识与可见度预测结合到检测框架中,以处理不同遮挡程度的行人。大量的实验表明,我们的VRGNet在加州理工-美国数据集的不同评估设置下都取得了领先的性能,特别是对于遮挡的行人。此外,与其他最先进的行人检测算法相比,该算法在CityPersons数据集的Heavy、Partial和Bare设置下分别达到了48.4%、9.3%和6.7%的竞争力,同时保持了更好的速度和准确性权衡。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信