{"title":"学习一种基于特征融合和上下文学习的判别特征用于目标检测","authors":"You Lei, Hongpeng Wang, Y. Wang","doi":"10.1109/SPAC.2017.8304337","DOIUrl":null,"url":null,"abstract":"Object detection is one of the most challenging tasks in the field of computer vision. It is widely used in traffic sign detection[1], pedestrian detection[2,3], person re-identification[4], object tracking[5,6,7] and so on[8,9]. Although convolutional neural network (CNN)-based algorithms have made great achievements in this field, object detection still suffers from illumination changes, occlusion, intraclass differences, etc.[10]. Candidate bounding box generation methods and feature extraction methods also influence the final detection result. In this paper, we propose a discriminative feature extraction method based on feature fusion and context learning.","PeriodicalId":161647,"journal":{"name":"2017 International Conference on Security, Pattern Analysis, and Cybernetics (SPAC)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Learning a discriminative feature for object detection based on feature fusing and context learning\",\"authors\":\"You Lei, Hongpeng Wang, Y. Wang\",\"doi\":\"10.1109/SPAC.2017.8304337\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Object detection is one of the most challenging tasks in the field of computer vision. It is widely used in traffic sign detection[1], pedestrian detection[2,3], person re-identification[4], object tracking[5,6,7] and so on[8,9]. Although convolutional neural network (CNN)-based algorithms have made great achievements in this field, object detection still suffers from illumination changes, occlusion, intraclass differences, etc.[10]. Candidate bounding box generation methods and feature extraction methods also influence the final detection result. In this paper, we propose a discriminative feature extraction method based on feature fusion and context learning.\",\"PeriodicalId\":161647,\"journal\":{\"name\":\"2017 International Conference on Security, Pattern Analysis, and Cybernetics (SPAC)\",\"volume\":\"39 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2017-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2017 International Conference on Security, Pattern Analysis, and Cybernetics (SPAC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SPAC.2017.8304337\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 International Conference on Security, Pattern Analysis, and Cybernetics (SPAC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SPAC.2017.8304337","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Learning a discriminative feature for object detection based on feature fusing and context learning
Object detection is one of the most challenging tasks in the field of computer vision. It is widely used in traffic sign detection[1], pedestrian detection[2,3], person re-identification[4], object tracking[5,6,7] and so on[8,9]. Although convolutional neural network (CNN)-based algorithms have made great achievements in this field, object detection still suffers from illumination changes, occlusion, intraclass differences, etc.[10]. Candidate bounding box generation methods and feature extraction methods also influence the final detection result. In this paper, we propose a discriminative feature extraction method based on feature fusion and context learning.