{"title":"Semi-Automatic Prediction of Landmarks on Human Models in Varying Poses","authors":"S. Wuhrer, Z. B. Azouz, Chang Shu","doi":"10.1109/CRV.2010.25","DOIUrl":"https://doi.org/10.1109/CRV.2010.25","url":null,"abstract":"We present an algorithm to predict landmarks on 3D human scans in varying poses. Our method is based on learning bending-invariant landmark properties. We also learn the spatial relationships between pairs of landmarks using canonical forms. The information is modeled by a Markov network, where each node of the network corresponds to a landmark position and where each edge of the network represents the spatial relationship between a pair of landmarks. We perform probabilistic inference over the Markov network to predict the landmark locations on human body scans in varying poses. We evaluated the algorithm on 200 models with different shapes and poses. The results show that most landmarks are predicted well.","PeriodicalId":358821,"journal":{"name":"2010 Canadian Conference on Computer and Robot Vision","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124822342","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
S. Metari, F. Prel, T. Moszkowicz, D. Laurendeau, N. Teasdale, S. Beauchemin
{"title":"A Computer Vision System for Analyzing and Interpreting the Cephalo-ocular Behavior of Drivers in a Simulated Driving Context","authors":"S. Metari, F. Prel, T. Moszkowicz, D. Laurendeau, N. Teasdale, S. Beauchemin","doi":"10.1109/CRV.2010.35","DOIUrl":"https://doi.org/10.1109/CRV.2010.35","url":null,"abstract":"In this paper we introduce a new computer vision framework for the analysis and interpretation of the cephalo-ocular behavior of drivers. We start by detecting the most important facial features, namely the nose tip and the eyes. For that, we introduce a new algorithm for eyes detection and we call upon the cascade of boosted classifiers technique based on Haar-like features for detecting the nose tip. Once those facial features are well identified, we apply the pyramidal Lucas-Kanade method for tracking purposes. Events resulting from those two approaches are combined in order to identify, analyze and interpret the cephalo-ocular behavior of drivers. Experimental results confirm both the robustness and the effectiveness of the proposed framework.","PeriodicalId":358821,"journal":{"name":"2010 Canadian Conference on Computer and Robot Vision","volume":"197 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121881472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Comparison of the EKF, SPKF, and the Bayes Filter for Landmark-Based Localization","authors":"Chi Hay Tong, T. Barfoot","doi":"10.1109/CRV.2010.33","DOIUrl":"https://doi.org/10.1109/CRV.2010.33","url":null,"abstract":"The conventional approach to nonlinear state estimation, the Extended Kalman Filter (EKF), is quantitatively compared to the performance of the relative newcomer, the Sigma-Point Kalman Filter (SPKF). These approaches are applied to the problem of localization of a mobile robot using a known map, and compared under the context of the practical best performance of a Bayes Filter-type method using a particle filter with a very large number of particles.","PeriodicalId":358821,"journal":{"name":"2010 Canadian Conference on Computer and Robot Vision","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114873419","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Decoupled Active Surface for Volumetric Image Segmentation","authors":"A. Mishra, P. Fieguth, David A Clausi","doi":"10.1109/CRV.2010.45","DOIUrl":"https://doi.org/10.1109/CRV.2010.45","url":null,"abstract":"Finding the surface of a volumetric 3D object is a fundamental problem in computer vision. Energy minimizing splines, such as active surfaces, have been used to carry out such tasks, evolving under the influence of internal and external energies until the model converges to a desired surface. The present deformable model based surface extraction techniques are computationally expensive and are generally unreliable in identifying the surfaces of noisy, high-curvature and cluttered 3D objects. This paper proposes a novel decoupled active surface (DAS) for identifying the surface of volumetric 3D objects. The proposed DAS introduces two novel aspects which leads to robust, efficient and accurate convergence. First, rather than a parameterized surface, which leads to difficulties with complex shapes and parameter singularities, the DAS uses a conforming triangular mesh to represent the surface. Second, motivated by earlier successes in two-dimensional segmentation, the DAS treats the two energy components separately and uses novel solution techniques to efficiently minimize the two energy terms separately. The performance of DAS in segmenting static 3D objects is presented using several natural and synthetic volumetric images, with excellent convergence results.","PeriodicalId":358821,"journal":{"name":"2010 Canadian Conference on Computer and Robot Vision","volume":"104 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125216072","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Deformable Object Segmentation and Contour Tracking in Image Sequences Using Unsupervised Networks","authors":"A. Crétu, E. Petriu, P. Payeur, Fouad F. Khalil","doi":"10.1109/CRV.2010.43","DOIUrl":"https://doi.org/10.1109/CRV.2010.43","url":null,"abstract":"The paper discusses a novel unsupervised learning approach for tracking deformable objects manipulated by a robotic hand in a series of images collected by a video camera. The object of interest is automatically segmented from the initial frame in the sequence. The segmentation is treated as clustering based on color information and spatial features and an unsupervised network is employed to cluster each pixel of the initial frame. Each pixel from the clustering results is then classified as either object of interest or background and the contour of the object is identified based on this classification. Using static (color) and dynamic (motion between frames) information, the contour is then tracked with an algorithm based on neural gas networks in the sequence of images. Experiments performed under different conditions reveal that the method tracks accurately the test objects even for severe contour deformations, is fast and insensitive to smooth changes in lighting, contrast and background.","PeriodicalId":358821,"journal":{"name":"2010 Canadian Conference on Computer and Robot Vision","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127658199","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Mammogram Image Superresolution Based on Statistical Moment Analysis","authors":"A. Wong, A. Mishra, David A Clausi, P. Fieguth","doi":"10.1109/CRV.2010.51","DOIUrl":"https://doi.org/10.1109/CRV.2010.51","url":null,"abstract":"A novel super resolution method for enhancing the resolution of mammogram images based on statistical moment analysis (SMA) has been designed and implemented. The proposed SMA method enables high resolution mammogram images to be produced at lower levels of radiation exposure to the patient. The SMA method takes advantage of the statistical characteristics of the underlying breast tissues being imaged to produce high resolution mammogram images with enhanced fine tissue details such that the presence of masses and micro calcifications can be more easily identified. In the SMA method, the super resolution problem is formulated as a constrained optimization problem using an adaptive third-order Markov prior model, and solved efficiently using a conjugate gradient approach. The priors are adapted based on the inter-pixel likelihoods of the first moment about zero (mean), second central moment (variance), and third and fourth standardized moments (skewness and kurtosis) from the low resolution images. Experimental results demonstrate the effectiveness of the SMA method at enhancing fine tissue details when compared to existing resolution enhancement methods.","PeriodicalId":358821,"journal":{"name":"2010 Canadian Conference on Computer and Robot Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128267948","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Automatic Position Registration of Street-level Fisheye Images into Aerial Image Using Line Structures and Mutual Information","authors":"M. Zouqi, J. Samarabandu, Yanbo Zhou","doi":"10.1109/CRV.2010.22","DOIUrl":"https://doi.org/10.1109/CRV.2010.22","url":null,"abstract":"Geospatial imaging is a relatively new term which is increasingly becoming more important for both government and commercial sectors. Images taken at street level can be geo-coded using a camera equipped with a built-in GPS device. However, the location that GPS provides are prone to errors up to 10 meters. In this paper we propose an algorithm to find the accurate location of a street-level image taken with a fisheye camera within a satellite image. Our algorithm is based on straight line detection and matching using Hough transform and gradient information around the detected lines. The rotation parameter is obtained using the best corresponding lines. Then mutual information (MI) is used as the similarity measure along the best match lines to determine the translational parameters. Moreover, as the correction process is carried out for a consecutive series of images rather than an individual image, the final location of each image will be assessed to be consistent with its neighboring images.","PeriodicalId":358821,"journal":{"name":"2010 Canadian Conference on Computer and Robot Vision","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130787463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Probabilistic Tracking of Pedestrian Movements via In-Floor Force Sensing","authors":"R. Rajalingham, Y. Visell, J. Cooperstock","doi":"10.1109/CRV.2010.26","DOIUrl":"https://doi.org/10.1109/CRV.2010.26","url":null,"abstract":"This article presents a probabilistic approach to the tracking and estimation of the lower body posture of users moving on foot over an instrumented floor surface. The latter consists of an array of low-cost force platforms providing intermittent foot-floor contact data with limited spatial resolution. We use this data to track body posture in 3D space using Bayesian filters with a switching state-space model. Potential applications of this work to person tracking and human-computer interaction are described.","PeriodicalId":358821,"journal":{"name":"2010 Canadian Conference on Computer and Robot Vision","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114221772","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Matching Maximally Stable Extremal Regions Using Edge Information and the Chamfer Distance Function","authors":"P. Elinas","doi":"10.1109/CRV.2010.10","DOIUrl":"https://doi.org/10.1109/CRV.2010.10","url":null,"abstract":"We consider the problem of image recognition using local features. We present a method for matching Maximally Stable Extremal Regions using edge information and the chamfer distance function. We represent MSERs using the Canny edges of their binary image representation in an affine normalized coordinate frame and find correspondences using chamfer matching. We evaluate the performance of our approach on a large number of data sets commonly used in the computer vision literature and we show that it is useful for matching images under large affine and viewpoint transformations as well as blurring, illumination changes and JPEG compression artifacts.","PeriodicalId":358821,"journal":{"name":"2010 Canadian Conference on Computer and Robot Vision","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125146711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Selecting and Commanding Individual Robots in a Multi-Robot System","authors":"Alex Couture-Beil, R. Vaughan, Greg Mori","doi":"10.1109/CRV.2010.28","DOIUrl":"https://doi.org/10.1109/CRV.2010.28","url":null,"abstract":"We present a novel real-time computer vision-based system for facilitating interactions between a single human and a multi-robot system: a user first selects an individual robot from a group of robots, by simply looking at it, and then commands the selected robot with a motion-based gesture. Robots estimate which robot the user is looking at by performing a distributed leader election based on the \"score\" of the detected frontal face.","PeriodicalId":358821,"journal":{"name":"2010 Canadian Conference on Computer and Robot Vision","volume":"7 2","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133135563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}