{"title":"Color Photometric Stereo for Albedo and Shape Reconstruction","authors":"O. Ikeda, Y. Duan","doi":"10.1109/WACV.2008.4544015","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544015","url":null,"abstract":"A new color photometric stereo method is presented, which reconstructs both albedo and shape from three color images. Optical polarization filters are placed in front of the light source and a camera to capture images free from specular reflection components. First, the albedo maps are derived from three color images. Then, the shape is reconstructed based on the Lambertian reflection from both the images and the albedo maps, using the Jacobi iterative method. The reconstruction is optimized using local and global weights, to minimize degradations due to shadows and saturations in images and color-dependent reflection characteristics of objects. Experimental results are given to evaluate the method.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124886158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Distributed Visual Processing for a Home Visual Sensor Network","authors":"Kwangsu Kim, G. Medioni","doi":"10.1109/WACV.2008.4544043","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544043","url":null,"abstract":"We address issues dealing with distributed visual processing for a personal service robot in the Intelligent Home environment. We propose an efficient and reliable framework to organize and coordinate the vision sensor nodes: fixed cameras mounted on walls, and camera(s) on the mobile robot. We also propose key visual functionalities necessary for the robot to perform its activities. They include people detection and identification, action recognition, gesture recognition, and self-localization. We propose solutions to the different vision tasks, and present our implementation within this framework, validated with experimental results.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129202716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Localization and Segmentation of A 2D High Capacity Color Barcode","authors":"Devi Parikh, Gavin Jancke","doi":"10.1109/WACV.2008.4544033","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544033","url":null,"abstract":"A 2D color barcode can hold much more information than a binary barcode. Barcodes are often intended for consumer use where using a cellphone, a consumer can take an image of a barcode on a product, and retrieve relevant information about the product. The barcode must be read using computer vision techniques. While a color barcode can hold more information, it makes this vision task in consumer scenarios unusually challenging. We present our approach to the localization and segmentation of a 2D color barcode in such challenging scenarios, along with its evaluation on a diverse collection of images of Microsoft's recently launched high capacity color barcode (HCCB). We exploit the unique trait of barcode reading: the barcode decoder can give the vision algorithm feedback, and develop a progressive strategy to achieve both - high accuracy in diverse scenarios as well as computational efficiency.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"231 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123732015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Recovering Social Networks From Massive Track Datasets","authors":"C. Connolly, J. Burns, H. Bui","doi":"10.1109/WACV.2008.4544042","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544042","url":null,"abstract":"Analysis of massive track datasets is a challenging problem, especially when examining n-way relations inherent in social networks. In this paper, we use the Mitsubishi track database to examine the usefulness of three types of interaction features observable in tracklet networks. We explore ways in which social network information can be extracted and visualized using a statistical sampling of these features from a very large track dataset, with very little ground truth or outside knowledge. Special attention is given to methods that are likely to scale well beyond the size of the Mitsubishi dataset.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114151222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Explanation-Based Object Recognition","authors":"Geoffrey Levine, G. DeJong","doi":"10.1109/WACV.2008.4544019","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544019","url":null,"abstract":"Many of today's visual scene and object categorization systems learn to classify using a statistical profile over a large number of small-scale local features sampled from the image. While some application systems have been constructed, this technology has enjoyed far more success in the research setting. The approach is best suited to tasks where within-class variability is small compared to between-class variability. This condition holds for large diverse artificial collections such as CalTech 101 where most categories have little to do with each other, but it often does not hold among naturalistic application-driven categories. Here, category distinctions are more likely to be conceptual or functional, and within-class differences can rival or exceed between- class differences. In this paper, we show how the local feature approach can be extended using explanation-based learning (EBL). The EBL approach makes use of readily available prior domain knowledge assembled into plausible explanations for why a training example's observable features might merit its assigned training label. Explanations expose additional semantic features and suggest how those hidden features may be estimated from observable features. We exhibit our approach on two CalTech 101 dataset tasks that we argue are emblematic of applied domains: Ketch vs. Schooner and Airplane vs. Background. In both cases classification accuracy is significantly improved.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132331339","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Cata-Fisheye Camera for Panoramic Imaging","authors":"Gurunandan Krishnan, S. Nayar","doi":"10.1109/WACV.2008.4544035","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544035","url":null,"abstract":"We present a novel panoramic imaging system which uses a curved mirror as a simple optical attachment to a fish- eye lens. When compared to existing panoramic cameras, our \"'cata-fisheye\" camera has a simple, compact and inexpensive design, and yet yields high optical performance. It captures the desired panoramic field of view in two parts. The upper part is obtained directly by the fisheye lens and the lower part after reflection by the curved mirror. These two parts of the field of view have a small overlap that is used to stitch them into a single seamless panorama. The cata-fisheye concept allows us to design cameras with a wide range of fields of view by simply varying the parameters and position of the curved mirror. We provide an automatic method for the one-time calibration needed to stitch the two parts of the panoramic field of view. We have done a complete performance evaluation of our concept with respect to (i) the optical quality of the captured images, (ii) the working range of the camera over which the parallax is negligible, and (iii) the spatial resolution of the computed panorama. Finally, we have built a prototype cata-fisheye video camera with a spherical mirror that can capture high resolution panoramic images (3600times550pixels) with a 360deg (horizontal) x 55deg (vertical) field of view.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132480322","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Recognition of Human Actions using an Optimal Control Based Motor Model","authors":"Sumitra Ganesh, R. Bajcsy","doi":"10.1109/WACV.2008.4544021","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544021","url":null,"abstract":"We present a novel approach to the problem of representation and recognition of human actions, that uses an optimal control based model to connect the high-level goals of a human subject to the low-level movement trajectories captured by a computer vision system. These models quantify the high-level goals as a performance criterion or cost function which the human sensorimotor system optimizes by picking the control strategy that achieves the best possible performance. We show that the human body can be modeled as a hybrid linear system that can operate in one of several possible modes, where each mode corresponds to a particular high-level goal or cost function. The problem of action recognition, then is to infer the current mode of the system from observations of the movement trajectory. We demonstrate our approach on 3D visual data of human arm motion.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"209 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133862017","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Locally Adjusted Robust Regression for Human Age Estimation","authors":"G. Guo, Yun Fu, Thomas S. Huang, C. Dyer","doi":"10.1109/WACV.2008.4544009","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544009","url":null,"abstract":"Automatic human age estimation has considerable potential applications in human computer interaction and multimedia communication. However, the age estimation problem is challenging. We design a locally adjusted robust regressor (LARR) for learning and prediction of human ages. The novel approach reduces the age estimation errors significantly over all previous methods. Experiments on two aging databases show the success of the proposed method for human aging estimation.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"393 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132626499","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Chao Zhang, P. Chockalingam, Ankit Kumar, P. Burt, Arvind Lakshmikumar
{"title":"Qualitative Assessment of Video Stabilization and Mosaicking Systems","authors":"Chao Zhang, P. Chockalingam, Ankit Kumar, P. Burt, Arvind Lakshmikumar","doi":"10.1109/WACV.2008.4544008","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544008","url":null,"abstract":"Image stabilization is a key preprocessing step in dynamic image analysis, which deals with the removal of unwanted motion in a video sequence. It is principally understood as the warping of video sequences resulting in a total or partial removal of image motion. Stabilization is invaluable for motion analysis, structure from motion, independent motion detection, geo-registration and mosaicking, autonomous vehicle navigation, model-based compression, and many others. Given the usefulness of image stabilization for many applications, a variety of algorithms have been proposed to perform this task, and many real-time systems have been built to stabilize the real-time videos and provide motion data for tracking and geo-registrations. However, even though there are on-line libraries that provide test videos, there has been no established methods or industrial standards based on which the performance of a stabilization algorithm or system can be measured. This paper aims to address this gap and suggests an evaluation methodology which would provide us the ability to qualitatively measure the performance of a given stabilized system. We propose a performance measurement system and define the performance metrics in this paper. We then apply the assessment to two typical stabilization systems. The discussed methods can be used to benchmark video stabilization systems.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130766368","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Online/Realtime Structure and Motion for General Camera Models","authors":"G. Schweighofer, Sinisa Segvic, A. Pinz","doi":"10.1109/WACV.2008.4544016","DOIUrl":"https://doi.org/10.1109/WACV.2008.4544016","url":null,"abstract":"This paper presents a novel algorithm for online structure and motion estimation. The algorithm works for general camera models and minimizes object space error, it does not rely on gradient-based optimization, and it is provably globally convergent. In comparison to previous work, which reports cubic complexity in the number of frames, our major contribution is a significant reduction of complexity. The new algorithm requires constant time per frame and can thus be used in online applications. Experimental results show high reconstruction accuracy with respect to simulated ground truth data. We also present two applications in artificial marker reconstruction and handheld augmented reality.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126676743","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}