{"title":"Ceiling analysis of pedestrian recognition pipeline for an autonomous car application","authors":"H. Roncancio, A. C. Hernandes, M. Becker","doi":"10.1109/WORV.2013.6521941","DOIUrl":"https://doi.org/10.1109/WORV.2013.6521941","url":null,"abstract":"This paper presents an exploration of the ceiling analysis of machine learning systems. It also provides an approach to the development of pedestrian recognition systems using this analysis. A pedestrian detection pipeline is simulated in order to evaluate this method. The advantage of this method is that it allows determining the most promising pipeline's elements to be modified as a way of more efficiently improving the recognition system. The pedestrian recognition is based on computer vision and is intended for an autonomous car application. A Linear SVM used as classifier enables the recognition, so this development is also addressed as a machine learning problem. This analysis concludes that for this application the more worthy path to be followed is the improvement of the pre-processing method instead of the classifier.","PeriodicalId":130461,"journal":{"name":"2013 IEEE Workshop on Robot Vision (WORV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129076908","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Color-based detection robust to varying illumination spectrum","authors":"M. Linderoth, A. Robertsson, Rolf Johansson","doi":"10.1109/WORV.2013.6521924","DOIUrl":"https://doi.org/10.1109/WORV.2013.6521924","url":null,"abstract":"In color-based detection methods, varying illumination often causes problems, since an object may be perceived to have different colors under different lighting conditions. In the field of color constancy this is usually handled by estimating the illumination spectrum and accounting for its effect on the perceived color. In this paper a method for designing a robust classifier is presented, i.e., instead of estimating and adapting to different lighting conditions, the classifier is made wider to detect a colored object for a given range of lighting conditions. This strategy also naturally handles the case where different parts of an object are illuminated by different light sources at the same time. Only one set of training data per light source has to be collected, and then the detector can handle any combination of the light sources for a large range of illumination intensities.","PeriodicalId":130461,"journal":{"name":"2013 IEEE Workshop on Robot Vision (WORV)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125477802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Detecting partially occluded objects via segmentation and validation","authors":"M. Levihn, M. Dutton, A. J. Trevor, M. Silman","doi":"10.1109/WORV.2013.6521925","DOIUrl":"https://doi.org/10.1109/WORV.2013.6521925","url":null,"abstract":"This paper presents a novel algorithm: Verfied Partial Object Detector (VPOD) for accurate detection of partially occluded objects such as furniture in 3D point clouds. VPOD is implemented and validated on real sensor data obtained by our robot. It extends Viewpoint Feature Histograms (VFH), which classify unoccluded objects, to also classify partially occluded objects such as furniture that might be seen in typical office environments. To achieve this result, VPOD employs two strategies. First, object models are segmented and the object database is extended to include partial models. Second, once a matching partial object is detected, the complete object model is aligned back into the scene and verified for consistency with the point cloud data. Overall, our approach increases the number of objects found and substantially reduces false positives due to the verification process.","PeriodicalId":130461,"journal":{"name":"2013 IEEE Workshop on Robot Vision (WORV)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127687352","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Shanshan Zhang, C. Bauckhage, D. A. Klein, A. Cremers
{"title":"Moving pedestrian detection based on motion segmentation","authors":"Shanshan Zhang, C. Bauckhage, D. A. Klein, A. Cremers","doi":"10.1109/WORV.2013.6521921","DOIUrl":"https://doi.org/10.1109/WORV.2013.6521921","url":null,"abstract":"The detection of moving pedestrians is of major importance in the area of robot vision, since information about such persons and their tracks should be incorporated into reliable collision avoidance algorithms. In this paper, we propose a new approach, based on motion analysis, to detect moving pedestrians. Our main contribution is to localize moving objects by motion segmentation on an optical flow field as a preprocessing step, so as to significantly reduce the number of detection windows needed to be evaluated by a subsequent people classifier, resulting in a fast method for real-time systems. Therefore, we align detection windows with segmented motion-blobs using a height-prior rule. Finally, we apply a Histogram of Oriented Gradients (HOG) features based Support Vector Machine with Radial Basis Function kernel (RBF-SVM) to estimate a confidence for each detection window, and thereby locate potential pedestrians inside the segmented blobs. Experimental results on “Daimler mono moving pedestrian detection” benchmark show that our approach obtains a log-average miss rate of 43% in the FPPI range [10-2, 100], which is a clear improvement with respect to the naive HOG+linSVM approach and better than several other state-of-the-art detectors. Moreover, our approach also reduces runtime per frame by an order of magnitude.","PeriodicalId":130461,"journal":{"name":"2013 IEEE Workshop on Robot Vision (WORV)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129532131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Human-object-object-interaction affordance","authors":"Shaogang Ren, Yu Sun","doi":"10.1109/WORV.2013.6521912","DOIUrl":"https://doi.org/10.1109/WORV.2013.6521912","url":null,"abstract":"This paper presents a novel human-object-object (HOO) interaction affordance learning approach that models the interaction motions between paired objects in a human-object-object way and use the motion models to improve the object recognition reliability. The innate interaction-affordance knowledge of the paired objects is modeled from a set of labeled training data that contains relative motions of the paired objects, humans actions, and object labels. The learned knowledge of the pair relationship is represented with a Bayesian Network and the trained network is used to improve recognition reliability of the objects.","PeriodicalId":130461,"journal":{"name":"2013 IEEE Workshop on Robot Vision (WORV)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121696083","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Panorama creation using a team of robots","authors":"Yongqiang Huang, W. Snyder","doi":"10.1109/WORV.2013.6521922","DOIUrl":"https://doi.org/10.1109/WORV.2013.6521922","url":null,"abstract":"A system is presented which allows a single human to teleoperate a team of camera-equipped robots. This paper emphasizes the image processing required to take a number of views and construct a single panorama which provides a sense of a 3-D environment to the operator who finds it easy to comprehend the environment and to control the team using something as simple as a joystick. Since the cameras have diverse poses, their output images must be distorted to provide smooth alignment. This is accomplished by correspondence finding, triangular tessellation and warping of a portion of each view. The panorama which gives a 180° field-of-view is projected onto a semi-circular array of monitors to provide the operator with a sensation of both forward and peripheral views.","PeriodicalId":130461,"journal":{"name":"2013 IEEE Workshop on Robot Vision (WORV)","volume":"146 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122242642","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Active object recognition using vocabulary trees","authors":"N. Govender, J. Claassens, F. Nicolls, J. Warrell","doi":"10.1109/WORV.2013.6521945","DOIUrl":"https://doi.org/10.1109/WORV.2013.6521945","url":null,"abstract":"For mobile robots to perform certain tasks in human environments, fast and accurate object classification is essential. Actively exploring objects by changing viewpoints promises an increase in the accuracy of object classification. This paper presents an efficient feature-based active vision system for the recognition and verification of objects that are occluded, appear in cluttered scenes and may be visually similar to other objects present. This system is designed using a selector-observer framework where the selector is responsible for the automatic selection of the next best viewpoint and a Bayesian `observer' updates the belief hypothesis and provides feedback. A new method for automatically selecting the `next best viewpoint' is presented using vocabulary trees. It is used to calculate a weighting for each feature based on its perceived uniqueness, allowing the system to select the viewpoint with the greatest number of `unique' features. The process is sped-up as new images are only captured at the `next best viewpoint' and processed when the belief hypothesis of an object is below some pre-defined threshold. The system also provides a certainty measure for the objects identity. This system out performs randomly selecting a viewpoint as it processes far fewer viewpoints to recognise and verify objects in a scene.","PeriodicalId":130461,"journal":{"name":"2013 IEEE Workshop on Robot Vision (WORV)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132019649","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Design of microassembly system and research on coarse-to-fine alignment strategy in combination with active zooming","authors":"Zhengtao Zhang, Juan Zhang, De Xu","doi":"10.1109/WORV.2013.6521917","DOIUrl":"https://doi.org/10.1109/WORV.2013.6521917","url":null,"abstract":"In this paper, a microassembly system based on 3-channles microvision system used for assembling 2 microparts named Si arm and Al shell is presented. Firstly, the structure of the system is described in detail. Then, a coarse-to-fine alignment strategy in combination with active zooming algorithm is presented. In the coarse alignment stage, alignment process is conducted with maximum field of view (FOV). In the fine alignment stage, the microscope is of maximum magnification to ensure the highest assembly accuracy. At last, the relative pose of the microparts is estimated in the assembly procedure. The experiment results show that the proposed algorithms can detect the assembly parameters online precisely.","PeriodicalId":130461,"journal":{"name":"2013 IEEE Workshop on Robot Vision (WORV)","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126975796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Incorporating local and global information using a novel distance function for scene recognition","authors":"E. Farahzadeh, Tat-Jen Cham, W. Li","doi":"10.1109/WORV.2013.6521927","DOIUrl":"https://doi.org/10.1109/WORV.2013.6521927","url":null,"abstract":"In the field of scene recognition using only one type of visual feature is not powerful enough to discriminate scene categories. In this paper we propose an innovative method to integrate global and local feature space into a map function based on a novel distance function. A subset of train images denoted as exemplar-set are selected. The local and global distances are defined according to the images in the exemplar-set. Distances are defined such that they indicate the contribution of different semantic aspects and global information in each scene category. An empirical study has been performed on the 15-Scene dataset in order to demonstrate the impact of appropriately incorporating both local and global information for the purpose of scene recognition. The experiments show, our model achieved state-of-the-art accuracy of 87.47.","PeriodicalId":130461,"journal":{"name":"2013 IEEE Workshop on Robot Vision (WORV)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128506785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Monocular visual odometry from frame to frame intensity differences for planetary exploration mobile robots","authors":"G. Martinez","doi":"10.1109/WORV.2013.6521914","DOIUrl":"https://doi.org/10.1109/WORV.2013.6521914","url":null,"abstract":"Traditionally stereo visual odometry algorithms estimate the robot's motion by maximizing the conditional probability of the 3D correspondences between two sets of 3D feature point positions, which were previously obtained from two consecutive stereo image pairs captured by a stereo video camera. As an alternative, in this paper a monocular visual odometry algorithm is proposed, which estimates the robot's motion by maximizing the conditional probability of the frame to frame intensity differences at observation points between two consecutive images captured by a monocular video camera. Experimental results with synthetic and real image sequences revealed highly accurate and reliable estimates, respectively. Additionally, it seems to be an excellent candidate for mobile robot missions where space, weight and power supply are really very limited.","PeriodicalId":130461,"journal":{"name":"2013 IEEE Workshop on Robot Vision (WORV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129854530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}