{"title":"A Best-Next-View-Selection Algorithm for Multi-view Rendering","authors":"Daniel Jung, R. Koch","doi":"10.1109/3DIMPVT.2011.49","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.49","url":null,"abstract":"In the last few years several multi-view display systems appeared increasing in size and the number of views that can be displayed. The increasing number of views allow for more realistic content with a high level of detail and complex lighting. For highest quality the images of synthetic scenes are usually ray traced. For large displays that are able to resolve a large amount of different views the number of viewing rays is enormous and it would take several month to render the content via ray tracing. One solution is to render only a few images and interpolate the remaining images with an image based rendering approach. This work introduces an algorithm that selects the best viewpoints for an image based rendering algorithm on the basis of given criteria. Special care is taken to cover occlusions and account for the dimension and angular resolution of multi-view displays.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130155939","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Xun Sun, Xing Mei, Shaohui Jiao, Mingcai Zhou, Haitao Wang
{"title":"Stereo Matching with Reliable Disparity Propagation","authors":"Xun Sun, Xing Mei, Shaohui Jiao, Mingcai Zhou, Haitao Wang","doi":"10.1109/3DIMPVT.2011.24","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.24","url":null,"abstract":"In this paper, we propose a novel propagation-based stereo matching algorithm. Starting from an initial disparity map, our algorithm selects highly reliable pixels and propagates their disparities along the scan line to produce dense disparity results. The key idea is to construct a line segment region for each pixel with local color and connectivity constraints. The pixel wise line segments are efficiently used to compute initial disparities, select reliable pixels and determine proper propagation regions. Streaking artifacts are effectively removed in a refinement process. Experimental results demonstrate the performance of the proposed method: it ranks 5th in the Middlebury benchmark, and the results can be computed within a few seconds.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122322559","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Pose Alignment for 3D Models and Single View Stereo Point Clouds Based on Stable Planes","authors":"A. Aldoma, M. Vincze","doi":"10.1109/3DIMPVT.2011.54","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.54","url":null,"abstract":"This paper presents a novel method for alignment of geometrically similar 3D models. It is based on the prior that both models have at least one common tangent plane on which both can stand stably and when standing on it the models are partially aligned. To determine the final rotation around the stable plane's normal, needed for a complete alignment, we adapt an image alignment technique based on the log-polar transformation. Because the set of stable planes of a model is small enough, alignment is efficiently approached as a global optimization problem that finds the common stable plane providing the best alignment according to a similarity measure. As the method does not rely on any kind of global symmetry features, we show it can be used to register incomplete stereo point clouds of objects located on a stable plane (table, ground, etc.) with the corresponding similar 3D models. We evaluate the 3D-alignment method by comparing it to the well-known CPCA and show a significant improvement when aligning 120 models belonging to 12 different classes.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133054303","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Façade Segmentation in a Multi-view Scenario","authors":"Michal Recky, Andreas Wendel, F. Leberl","doi":"10.1109/3DIMPVT.2011.52","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.52","url":null,"abstract":"We examine a new method of façade segmentation in a multi-view scenario. A set of overlapping, thus redundant street-side images exists and each image shows multiple buildings. A semantic segmentation identifies primary areas in the image such as sky, ground, vegetation, and façade. Subsequently, repeated patterns are detected in image segments previous labeled as \"façade areas\" and are applied to separate specific facades from each other. Experimentation is based on an industrial street-view dataset from a moving car by well-designed, calibrated, automated cameras. High overlap images define a multi-view scenario. We achieve 97% pixel-wise segmentation effectiveness, outperforming current state-of-the-art methods.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116157427","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"3D Data Segmentation by Local Classification and Markov Random Fields","authors":"Federico Tombari, L. D. Stefano","doi":"10.1109/3DIMPVT.2011.34","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.34","url":null,"abstract":"Object segmentation in 3D data such as 3D meshes and range maps is an emerging topic attracting increasing research interest. This work proposes a novel method to perform segmentation relying on the use of 3D features. The deployment of a specific grouping algorithm based on a Markov Random Field model successively to classification allows at the same time yielding automatic segmentation of 3D data as well as deploying non-linear classifiers that can well adapt to the data characteristics. Moreover, we embed our approach in a framework that jointly exploits shape and texture information to improve the outcome of the segmentation stage. In addition to quantitative results on several 3D and 2.5D scenes, we also demonstrate the effectiveness of our approach on an online framework based on a stereo sensor.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124371269","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An Evaluation of Volumetric Interest Points","authors":"Tsz-Ho Yu, Oliver J. Woodford, R. Cipolla","doi":"10.1109/3DIMPVT.2011.62","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.62","url":null,"abstract":"This paper presents the first performance evaluation of interest points on scalar volumetric data. Such data encodes 3D shape, a fundamental property of objects. The use of another such property, texture (ie 2D surface colouration), or appearance, for object detection, recognition and registration has been well studied, 3D shape less so. However, the increasing prevalence of depth sensors and the diminishing returns to be had from appearance alone have seen a surge in shape-based methods. In this work we investigate the performance of several detectors of interest points in volumetric data, in terms of repeatability, number and nature of interest points. Such methods form the first step in many shape-based applications. Our detailed comparison, with both quantitative and qualitative measures on synthetic and real 3D data, both point-based and volumetric, aids readers in selecting a method suitable for their application.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122051212","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Skull Assembly and Completion Using Template-Based Surface Matching","authors":"Li Wei, Wei Yu, Maoqing Li, Xin Li","doi":"10.1109/3DIMPVT.2011.59","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.59","url":null,"abstract":"We present a skull assembly and completion framework based on shape matching. In order to assemble fragmented skulls, we need to compute rigid transformations from these fragments to their assembled geometry. We develop a reliable assembly pipeline where each fragment is matched and transformed to be aligned with the template. In order to further complete the assembled skull with several damaged regions, we use the template to repair damaged regions on the assembled skull. The entire pipeline has been conducted on several real skull models and demonstrated great robustness and effectiveness.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"348 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133181662","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Integrating LIDAR into Stereo for Fast and Improved Disparity Computation","authors":"H. Badino, Daniel F. Huber, T. Kanade","doi":"10.1109/3DIMPVT.2011.58","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.58","url":null,"abstract":"The fusion of stereo and laser range finders (LIDARs) has been proposed as a method to compensate for each individual sensor's deficiencies - stereo output is dense, but noisy for large distances, while LIDAR is more accurate, but sparse. However, stereo usually performs poorly on textureless areas and on scenes containing repetitive structures, and the subsequent fusion with LIDAR leads to a degraded estimation of the 3D structure. In this paper, we propose to integrate LIDAR data directly into the stereo algorithm to reduce false positives while increasing the density of the resulting disparity image on textureless regions. We demonstrate with extensive experimental results with real data that the disparity estimation is substantially improved while speeding up the stereo computation by as much as a factor of five.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"517 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133564888","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Dynamic Shape Capture via Periodical-Illumination Optical Flow Estimation and Multi-view Photometric Stereo","authors":"Ying Fu, Yebin Liu, Qionghai Dai","doi":"10.1109/3DIMPVT.2011.12","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.12","url":null,"abstract":"Multi-view photometric stereo is well established for the shape recovery of static objects. However, it is difficult to align motion images under varying illumination so as to perform photometric stereo reconstruction for dynamic objects. To tackle this issue, this paper presents an optical flow estimation approach which works under periodically varying illuminations, and in cooperation with photometric stereo, enables high-quality 3D reconstruction of dynamic objects. Firstly, multi-view images of the moving object are captured under periodically varying illumination by the multi-camera multi-light system. Then, the optical flow is estimated to synthesize images under different illuminations for each viewpoint. Finally, the multi-view photometric stereo technique is employed to get a high accurate 3D model for each time instant. Experimental results on motion actors demonstrate that temporal successive images under varying illuminations are effectively registered, permitting accurate photometric reconstruction for moving objects.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122135044","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"3D Human Action Recognition for Multi-view Camera Systems","authors":"M. B. Holte, T. Moeslund, N. Nikolaidis, I. Pitas","doi":"10.1109/3DIMPVT.2011.50","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.50","url":null,"abstract":"This paper presents a novel approach for combining optical flow into enhanced 3D motion vector fields for human action recognition. Our approach detects motion of the actors by computing optical flow in video data captured by a multi-view camera setup with an arbitrary number of views. Optical flow is estimated in each view and extended to 3D using 3D reconstructions of the actors and pixel-to-vertex correspondences. The resulting 3D optical flow for each view is combined into a 3D motion vector field by taking the significance of local motion and its reliability into account. 3D Motion Context (3D-MC) and Harmonic Motion Context (HMC) are used to represent the extracted 3D motion vector fields efficiently and in a view-invariant manner, while considering difference in anthropometry of the actors and their movement style variations. The resulting 3D-MC and HMC descriptors are classified into a set of human actions using normalized correlation, taking into account the performing speed variations of different actors. We compare the performance of the 3D-MC and HMC descriptors, and show promising experimental results for the publicly available i3DPost Multi View Human Action Dataset.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124717853","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}