{"title":"Human Pose Estimation with Rotated Geometric Blur","authors":"Bo Chen, N. Nguyen, Greg Mori","doi":"10.1109/WACV.2008.4544022","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544022","url":null,"abstract":"We consider the problem of estimating the pose of a human figure in a single image. Our method uses an exemplar-matching framework, where a test image is matched to a database of exemplars upon which body joint positions have been marked. We find the best matching exemplar for a test image by employing a variant of an existing deformable template matching framework. A hierarchical correspondence process is developed to improve the efficiency of the existing framework. Quantitative results on the CMUMoBo dataset verify the effectiveness of our approach.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126840906","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Segmentation of Salient Regions in Outdoor Scenes Using Imagery and 3-D Data","authors":"Gunhee Kim, Daniel F. Huber, M. Hebert","doi":"10.1109/WACV.2008.4544014","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544014","url":null,"abstract":"This paper describes a segmentation method for extracting salient regions in outdoor scenes using both 3-D laser scans and imagery information. Our approach is a bottom- up attentive process without any high-level priors, models, or learning. As a mid-level vision task, it is not only robust against noise and outliers but it also provides valuable information for other high-level tasks in the form of optimal segments and their ranked saliency. In this paper, we propose a new saliency definition for 3-D point clouds and we incorporate it with saliency features from color information.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121165863","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multi-Pose Face Detection with Asymmetric Haar Features","authors":"Geovany A. Ramírez, O. Fuentes","doi":"10.1109/WACV.2008.4544013","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544013","url":null,"abstract":"In this paper we present a system for multi-pose face detection. Our system presents three main contributions. First, we introduce the use of asymmetric Haar features. Asymmetric Haar features provide a rich feature space, which allows to build classifiers that are accurate and much simpler than those obtained with other features. The second contribution is the use of a genetic algorithm to search efficiently in the extremely large parameter space of potential features. Using this genetic algorithm, we generate a feature set that allows to exploit the expressive advantage of asymmetric Haar features and is small enough to permit exhaustive evaluation. The third contribution is the application of a skin color-segmentation scheme to reduce the search space. Our system uses specialized detectors in different face poses that are built using AdaBoost and the C4.5 rule induction algorithm. Experimental results using the CMU profile test set and BioID frontal faces test set, in addition to our own multi-pose face test set, show that our system is competitive with other systems presented recently in the literature.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133865260","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Charalambos (Charis) Poullis, Suya You, U. Neumann
{"title":"A Vision-Based System For Automatic Detection and Extraction Of Road Networks","authors":"Charalambos (Charis) Poullis, Suya You, U. Neumann","doi":"10.1109/WACV.2008.4543996","DOIUrl":"https://doi.org/10.1109/WACV.2008.4543996","url":null,"abstract":"In this paper we present a novel vision-based system for automatic detection and extraction of complex road networks from various sensor resources such as aerial photographs, satellite images, and LiDAR. Uniquely, the proposed system is an integrated solution that merges the power of perceptual grouping theory (Gabor filtering, tensor voting) and optimized segmentation techniques (global optimization using graph-cuts) into a unified framework to address the challenging problems of geospatial feature detection and classification. Firstly, the local precision of the Gabor filters is combined with the global context of the tensor voting to produce accurate classification of the geospatial features. In addition, the tensorial representation used for the encoding of the data eliminates the need for any thresholds, therefore removing any data dependencies. Secondly, a novel orientation-based segmentation is presented which incorporates the classification of the perceptual grouping, and results in segmentations with better defined boundaries and continuous linear segments. Finally, a set of Gaussian-based filters are applied to automatically extract centerline information (magnitude, width and orientation). This information is then used for creating road segments and then transforming them to their polygonal representations.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"164 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132136838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Xiaojin Gong, Bin Xu, C. Reed, C. Wyatt, D. Stilwell
{"title":"Real-time Robust Mapping for an Autonomous Surface Vehicle using an Omnidirectional Camera","authors":"Xiaojin Gong, Bin Xu, C. Reed, C. Wyatt, D. Stilwell","doi":"10.1109/WACV.2008.4544024","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544024","url":null,"abstract":"Towards the goal of achieving truly autonomous navigation for a surface vehicle in maritime environments, a critical task is to detect surrounding obstacles such as the shore, docks, and other boats. In this paper, we demonstrate a real-time vision-based mapping system which detects and localizes stationary obstacles using a single omnidirectional camera and navigational sensors (GPS and gyro). The main challenge of this work is to make mapping robust to a large number of outliers, which stem from waves and specular reflections on the surface of the water. To address this problem, a two-step robust outlier rejection method is proposed. Experimental results obtained in unstructured large-scale environments are presented and validated using topographic maps.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129943877","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
J. Tu, Yun Fu, A. Ivanovic, Thomas S. Huang, Li Fei-Fei
{"title":"Variational Transform Invariant Mixture of Probabilistic PCA","authors":"J. Tu, Yun Fu, A. Ivanovic, Thomas S. Huang, Li Fei-Fei","doi":"10.1109/WACV.2008.4543995","DOIUrl":"https://doi.org/10.1109/WACV.2008.4543995","url":null,"abstract":"In many video-based object recognition applications, the object appearances are acquired by visual tracking or detection and are inconsistent due to misalignments. We believe the misalignments can be removed if we can reduce the inconsistency in the object appearances caused by misalignments through clustering the objects in appearance, space and time domain simultaneously. We therefore propose to learn Transform Invariant Mixtures of Probabilistic PCA (TIMPPCA) model from the data while at the same time eliminating the misalignments. The model is formulated in a generative framework, and the misalignments are considered as hidden variables in the model. Variational EM update rules are then derived based on Variational Message Passing (VMP) techniques. The proposed TIMP-PCA is applied to improve head pose estimation performance and to detect the change of attention focus in meeting room video for meeting room video indexing/retrieval and achieves promising performance.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131635743","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Channel Segmentation using Confidence and Curvature-Guided Level Sets on Noisy Seismic Images","authors":"B. Kadlec","doi":"10.1109/WACV.2008.4544012","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544012","url":null,"abstract":"This paper presents a new method for segmenting channel features from commonly noisy 3D seismic images. Anisotropic diffusion using Gaussian-smoothed first order structure tensors is conducted along the strata of seismic images in a way that filters across discontinuous regions from noise or faulting, while preserving channel edges. The eigenstructure of the second order structure tensor is used to generate an estimation of orientation and channel curvature. Gaussian smoothing of second order tensor orientations accounts for noisy vectors from imprecise finite difference calculations and generates a stable tensor across the image. Analysis of the confidence and direction of second order eigenvectors is used to enhance depositional curvature in channel features by generating a confidence and curvature attribute. The tensor-derived attribute forms the terms of a PDE, which is iteratively updated as an implicit surface using the level set process. This technique is tested on two 3D seismic images with results that demonstrate the effectiveness of the approach.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124895287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Aniruddha Kembhavi, Ryan Farrell, Yuancheng Luo, D. Jacobs, R. Duraiswami, L. Davis
{"title":"Tracking Down Under: Following the Satin Bowerbird","authors":"Aniruddha Kembhavi, Ryan Farrell, Yuancheng Luo, D. Jacobs, R. Duraiswami, L. Davis","doi":"10.1109/WACV.2008.4544004","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544004","url":null,"abstract":"Socio biologists collect huge volumes of video to study animal behavior (our collaborators work with 30,000 hours of video). The scale of these datasets demands the development of automated video analysis tools. Detecting and tracking animals is a critical first step in this process. However, off-the-shelf methods prove incapable of handling videos characterized by poor quality, drastic illumination changes, non-stationary scenery and foreground objects that become motionless for long stretches of time. We improve on existing approaches by taking advantage of specific aspects of this problem: by using information from the entire video we are able to find animals that become motionless for long intervals of time; we make robust decisions based on regional features; for different parts of the image, we tailor the selection of model features, choosing the features most helpful in differentiating the target animal from the background in that part of the image. We evaluate our method, achieving almost 83% tracking accuracy on a more than 200,000 frame dataset of Satin Bowerbird courtship videos.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134221537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Location-based Services using Image Search","authors":"Pieter-Paulus Vertongen, D. Hansen","doi":"10.1109/WACV.2008.4544005","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544005","url":null,"abstract":"Recent developments in image search has made them sufficiently efficient to be used in real-time applications. GPS has become a popular navigation tool. While GPS information provide reasonably good accuracy, they are not always present in all hand held devices nor are they accurate in all situations, for example in urban environments. We propose a system to provide location-based services using image searches without requiring GPS. The goal of this system is to assist tourists in cities with additional information using their mobile phones and built-in cameras. Based upon the result of the image search engine and database image location knowledge, the location is determined of the query image and associated data can be presented to the user.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133931948","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Image Rendering Based on a Spatial Extension of the CIECAM02","authors":"Olivier Tulet, M. Larabi, C. Fernandez-Maloigne","doi":"10.1109/WACV.2008.4544030","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544030","url":null,"abstract":"With the multiplicity of imaging devices, the color quality and portability have become a very challenging problem. Moreover, a color is perceived with regards to its environment. So, if this environment changes it implies a change in the perceived color. In order to address this influence, the CIE (Commission Internationale de I'eclairage) has standardized a tool named color appearance model (CIECAM97*, CIECAM02). These models are able to take into account many phenomena related to human vision of color and can predict the color of a stimulus, function of its observations conditions. However, these models do not deal with the influence of spatial frequencies which can have a big impact on our perception. In this paper, we present an extended version of the CIECAM02 that integrates a spatial model correcting the color in relation to its spatial frequency. Moreover, the previous model has been modified to deal with images and not only single stimulus. The main difference with the rendering models (e.g. iCAM) lies in the fact that the proposed model, takes into account the spatial repartition of a pixel in addition to its environment. The obtained results are sound and demonstrate the efficiency of the proposed extension. This has been checked thanks to a psychophysical study where observers were assigned the task of assessing the quality of the improved version in comparison to the original.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125432989","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}