Michael Weinmann, Christopher Schwartz, R. Ruiters, R. Klein
{"title":"A Multi-camera, Multi-projector Super-Resolution Framework for Structured Light","authors":"Michael Weinmann, Christopher Schwartz, R. Ruiters, R. Klein","doi":"10.1109/3DIMPVT.2011.57","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.57","url":null,"abstract":"In this work, we present a framework for multi-camera, multi-projector object acquisition based on structured light. This approach allows the reconstruction of an object without moving either the object or the acquisition setup, avoiding any registration of independent measurements. To overcome the resolution limitations of the individual projectors, we introduce a novel super-resolution scheme. By exploiting high dynamic range imaging, we are able to handle even complicated objects, exhibiting strong specularities. We show that, combined with an iterated bundle adjustment, these improvements increase the accuracy of the obtained point cloud.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115690651","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Feature-Preserved Canonical Form for Non-rigid 3D Meshes","authors":"Z. Lian, A. Godil","doi":"10.1109/3DIMPVT.2011.22","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.22","url":null,"abstract":"Measuring the dissimilarity between non-rigid objects is a challenging problem in 3D shape retrieval. One potential solution is to construct the models' 3D canonical forms (i.e., isometry-invariant representations in 3D Euclidean space) on which any rigid shape matching algorithm can be applied. However, existing methods, which are typically based on embedding procedures, result in greatly distorted canonical forms, and thus could not provide satisfactory performance to distinguish non-rigid models. In this paper, we present a feature-preserved canonical form for non-rigid 3D meshes. The basic idea is to naturally deform original models against corresponding initial canonical forms calculated by Multidimensional Scaling (MDS). Specifically, objects are first segmented into near-rigid subparts, and then, through properly-designed rotations and translations, original subparts are transformed into poses that correspond well with their positions and directions on MDS canonical forms. Final results are obtained by solving some nonlinear minimization problems for optimal alignments and smoothing boundaries between subparts. Experiments on a widely utilized non-rigid 3D shape benchmark not only verify the advantages of our algorithm against existing approaches, but also demonstrate that, with the help of the proposed canonical form, we can obtain significantly better retrieval accuracy compared to the state-of-the-art.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129496015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Scene Cut: Class-Specific Object Detection and Segmentation in 3D Scenes","authors":"Jan Knopp, Mukta Prasad, L. Gool","doi":"10.1109/3DIMPVT.2011.30","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.30","url":null,"abstract":"In this paper we present a method to combine the detection and segmentation of object categories from 3D scenes. In the process, we combine the top-down cues available from object detection technique of Implicit Shape Models and the bottom-up power of Markov Random Fields for the purpose of segmentation. While such approaches have been tried for the 2D image problem domain before, this is the first application of such a method in 3D. 3D scene understanding is prone to many problems different from 2D owing to problems from noise, lack of distinctive high-frequency feature information, mesh parametrization problems etc. Our method enables us to localize objects of interest for more purposeful meshing and subsequent scene understanding.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128405003","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Cutting-Plane Training of Non-associative Markov Network for 3D Point Cloud Segmentation","authors":"Roman Shapovalov, A. Velizhev","doi":"10.1109/3DIMPVT.2011.10","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.10","url":null,"abstract":"We address the problem of object class segmentation of 3D point clouds. Each point of a cloud should be assigned a class label determined by the category of the object it belongs to. Non-associative Markov networks have been applied to this task recently. Indeed, they impose more flexible constraints on segmentation results in contrast to the associative ones. We show how to train non-associative Markov networks in a principled manner using the structured Support Vector Machine (SVM) formalism. In contrast to prior work we use the kernel trick which makes our method one of the first non-linear methods for max-margin Markov Random Field training applied to 3D point cloud segmentation. We evaluate our method on airborne and terrestrial laser scans. In comparison to the other non-linear training techniques our method shows higher accuracy.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132224105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Fast Stereo Matching of Feature Links","authors":"Chang-Il Kim, Soon-Yong Park","doi":"10.1109/3DIMPVT.2011.41","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.41","url":null,"abstract":"In stereo vision researches, feature-based stereo matching algorithms have been widely used in the preference of low computation cost and high matching accuracy. This paper presents a new stereo matching algorithm based on feature links. The proposed method, which is called feature link matching, utilizes the length and color information of feature links in stereo images. The proposed algorithm is very effective to decide correct correspondence, thus increases accuracy of stereo matching. In addition, inner features which lie within a link are interpolated by an internal division method to increase the number of correct disparity values. For real-time applications of the proposed method, point features are determined by the FAST extractor. Three feature link constraints, epipolar, ordering, and length, are employed. In experimental results, feature link matching yields 1 pixel disparity accuracy of 98.6% which is the average of 5 sample images from the Middlebury stereo data. Average computation time is about 18.7ms. The proposed matching algorithm is also applied to real-time 3D map building using a mobile robot.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"104 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123216083","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Automatic Keypoint Detection on 3D Faces Using a Dictionary of Local Shapes","authors":"Clement Creusot, Nick E. Pears, J. Austin","doi":"10.1109/3DIMPVT.2011.33","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.33","url":null,"abstract":"Keypoints on 3D surfaces are points that can be extracted repeatably over a wide range of 3D imaging conditions. They are used in many 3D shape processing applications, for example, to establish a set of initial correspondences across a pair of surfaces to be matched. Typically, keypoints are extracted using extremal values of a function over the 3D surface, such as the descriptor map for Gaussian curvature. That approach works well for salient points, such as the nose-tip, but can not be used with other less pronounced local shapes. In this paper, we present an automatic method to detect keypoints on 3D faces, where these keypoints are locally similar to a set of previously learnt shapes, constituting a 'local shape dictionary'. The local shapes are learnt at a set of 14 manually-placed landmark positions on the human face. Local shapes are characterised by a set of 10 shape descriptors computed over a range of scales. For each landmark, the proportion of face meshes that have an associated keypoint detection is used as a performance indicator. Repeatability of the extracted keypoints is measured across the FRGC v2 database.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"128 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134085104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Visual Hull from Imprecise Polyhedral Scene","authors":"Peng He, A. Edalat","doi":"10.1109/3DIMPVT.2011.28","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.28","url":null,"abstract":"We present a framework to compute the visual hull of a polyhedral scene, in which the vertices of the polyhedra are given with some imprecision. Two kinds of visual event surfaces, namely VE and EEE surfaces are modelled under the geometric framework to derive their counterpart object, namely partial VE and partial EEE surfaces, which contain the exact information of all possible visual event surfaces given the imprecision in the input. Correspondingly, a new definition of visual number is proposed to label the cells of Euclidean space partitioned by partial VE and partial EEE surfaces. The overall algorithm maintains the same computational complexity as the classical method and generates a partial visual hull which converges to the classical visual hull as the input converges to an exact value.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125203841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Toward Automatic 3D Generic Object Modeling from One Single Image","authors":"Min Sun, S. Kumar, G. Bradski, S. Savarese","doi":"10.1109/3DIMPVT.2011.11","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.11","url":null,"abstract":"We present a novel method for solving the challenging problem of generating 3D models of generic object categories from just one single un-calibrated image. Our method leverages the algorithm proposed in [1] which enables a partial reconstruction of the object from a single view. A full reconstruction is achieved in a subsequent object completion stage where modified or state-of-the-art 3D shape and texture completion techniques are used to recover the complete 3D model. We present results of our method on a number of images containing objects from five generic categories (mice, staplers, mugs, cars, and bicycles). We demonstrate (numerically and qualitatively) that our method produces convincing 3D models from a single image using minimal or no human intervention. Our technique is targeted to applications where users are interested in building virtual collections of 3D models of objects, and sharing such models in virtual environments such as Google 3D Warehouse or Second Life (secondlife.com).","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117245854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Sampling Relevant Points for Surface Registration","authors":"A. Torsello, E. Rodolà, A. Albarelli","doi":"10.1109/3DIMPVT.2011.43","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.43","url":null,"abstract":"Surface registration is a fundamental step in the reconstruction of three-dimensional objects. This is typically a two-step process where an initial coarse motion estimation is followed by a refinement step that almost invariably is some variant of Iterative Closest Point (ICP), which iteratively minimizes a distance function measured between pairs of selected neighboring points. The selection of relevant points on one surface to match against points on the other surface is an important issue in any efficient implementation of ICP, with strong implications both on the convergence speed and on the quality of the final alignment. This is due to the fact that typically on a surface there are a lot of low-curvature points that scarcely constrain the rigid transformation and an order of magnitude less descriptive points that are more relevant for finding the correct alignment. This results in a tendency of surfaces to \"over fit'' noise on low-curvature areas sliding away from the correct alignment. In this paper we propose a novel relevant-point sampling approach for ICP based on the idea that points in an area of great change constrain the transformation more and thus should be sampled with higher frequency. Experimental evaluations confront the alignment accuracy obtained with the proposed approach with those obtained with the commonly adopted uniform sub sampling and normal-space sampling strategies.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133289283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ryo Hanai, Kimitoshi Yamazaki, Hiroaki Yaguchi, K. Okada, M. Inaba
{"title":"Electric Appliance Parts Classification Using a Measure Combining the Whole Shape and Local Shape Distribution Similarities","authors":"Ryo Hanai, Kimitoshi Yamazaki, Hiroaki Yaguchi, K. Okada, M. Inaba","doi":"10.1109/3DIMPVT.2011.44","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.44","url":null,"abstract":"Classification of electric appliance parts is one of the interesting and practically valuable applications for 3D object recognition. Based on existing works, in this paper we try classifying electric appliance parts data obtained in an automatable process, which becomes a basis for automated recycling system. The dataset includes deformable objects such as cables as well as various rigid objects, some of which lacking a large part of the surface because of self-occlusions and materials of the parts. To realize high accuracy in classification, after the comparison of several similarity measures, we combine a measure which describes well the whole shape similarity with a measure that expresses the ratio of local surface patterns that appears in each model. The latter measure is suitable to describe the similarity of deformable objects that the whole shapes are heavily dependent on their configurations. We also investigate how the scale of computing local feature affects the classification result.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"59 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127026308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}