Zexia Liu, Chongyang Zhang, Yan Luo, Kai Chen, Qiping Zhou, Yunyu Lai
{"title":"利用知情环境改进小规模行人检测","authors":"Zexia Liu, Chongyang Zhang, Yan Luo, Kai Chen, Qiping Zhou, Yunyu Lai","doi":"10.1109/VCIP47243.2019.8965786","DOIUrl":null,"url":null,"abstract":"Finding small objects is fundamentally challenging because there is little signal on the object to exploit. For the small-scale pedestrian detection, one must use image evidence beyond the pedestrian extent, which is often formulated as context. Unlike existing object detection methods that use adjacent regions or whole image as the context simply, we focus on more informed contexts exploiting and utilizing to improve small-scale pedestrian detection: firstly, one relationship network is developed to utilize the correlation among pedestrian instances in one image; secondly, two spatial regions, overhead area and feet bottom area, are taken as spatial context to exploit the relevance between pedestrian and scenes; at last, GRU [7] (Gated Recurrent Units) modules are introduced to take encoded contexts as input to guide the feature selection and fusion of each proposal. Instead of getting all of the outputs at once, we also iterate twice to refine the detection incrementally. Comprehensive experiments on Caltech Pedestrian [8] and SJTU-SPID [9] datasets, indicate that, with more informed context, the detection performance can be improved significantly, especially for the small-scale pedestrians.","PeriodicalId":388109,"journal":{"name":"2019 IEEE Visual Communications and Image Processing (VCIP)","volume":"354 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Improving Small-Scale Pedestrian Detection Using Informed Context\",\"authors\":\"Zexia Liu, Chongyang Zhang, Yan Luo, Kai Chen, Qiping Zhou, Yunyu Lai\",\"doi\":\"10.1109/VCIP47243.2019.8965786\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Finding small objects is fundamentally challenging because there is little signal on the object to exploit. For the small-scale pedestrian detection, one must use image evidence beyond the pedestrian extent, which is often formulated as context. Unlike existing object detection methods that use adjacent regions or whole image as the context simply, we focus on more informed contexts exploiting and utilizing to improve small-scale pedestrian detection: firstly, one relationship network is developed to utilize the correlation among pedestrian instances in one image; secondly, two spatial regions, overhead area and feet bottom area, are taken as spatial context to exploit the relevance between pedestrian and scenes; at last, GRU [7] (Gated Recurrent Units) modules are introduced to take encoded contexts as input to guide the feature selection and fusion of each proposal. Instead of getting all of the outputs at once, we also iterate twice to refine the detection incrementally. Comprehensive experiments on Caltech Pedestrian [8] and SJTU-SPID [9] datasets, indicate that, with more informed context, the detection performance can be improved significantly, especially for the small-scale pedestrians.\",\"PeriodicalId\":388109,\"journal\":{\"name\":\"2019 IEEE Visual Communications and Image Processing (VCIP)\",\"volume\":\"354 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 IEEE Visual Communications and Image Processing (VCIP)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/VCIP47243.2019.8965786\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE Visual Communications and Image Processing (VCIP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/VCIP47243.2019.8965786","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Improving Small-Scale Pedestrian Detection Using Informed Context
Finding small objects is fundamentally challenging because there is little signal on the object to exploit. For the small-scale pedestrian detection, one must use image evidence beyond the pedestrian extent, which is often formulated as context. Unlike existing object detection methods that use adjacent regions or whole image as the context simply, we focus on more informed contexts exploiting and utilizing to improve small-scale pedestrian detection: firstly, one relationship network is developed to utilize the correlation among pedestrian instances in one image; secondly, two spatial regions, overhead area and feet bottom area, are taken as spatial context to exploit the relevance between pedestrian and scenes; at last, GRU [7] (Gated Recurrent Units) modules are introduced to take encoded contexts as input to guide the feature selection and fusion of each proposal. Instead of getting all of the outputs at once, we also iterate twice to refine the detection incrementally. Comprehensive experiments on Caltech Pedestrian [8] and SJTU-SPID [9] datasets, indicate that, with more informed context, the detection performance can be improved significantly, especially for the small-scale pedestrians.