2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission最新文献

筛选
英文 中文
Visual Words for 3D Reconstruction and Pose Computation 用于三维重建和姿态计算的视觉词
S. K., M. Berger, F. Sur
{"title":"Visual Words for 3D Reconstruction and Pose Computation","authors":"S. K., M. Berger, F. Sur","doi":"10.1109/3DIMPVT.2011.48","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.48","url":null,"abstract":"Visual vocabularies are standard tools in the object/image classification literature, and are emerging as a new tool for building point correspondences for pose estimation. This paper proposes several visual word based methods for point matching, with structure from motion and pose estimation applications in view. The three dimensional geometry of a scene is first extracted with bundle adjustment techniques based on the key point correspondences. These correspondences are obtained by grouping the set of all SIFT descriptors from the training images into visual words. We obtain a more accurate 3D geometry than with classical image-to-image point matching. In the second step, these visual words serve as 3D point descriptors robust to viewpoint change, and are then used for building 2D-3D correspondences for a test image, yielding the pose of the camera by solving the PnP problem. We compare several visual word formation techniques w.r.t robustness to viewpoint change between the learning and test images and discuss the required computational time.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124930661","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Real Time Detection of Repeated Structures in Point Clouds of Urban Scenes 城市场景点云中重复结构的实时检测
Sam Friedman, I. Stamos
{"title":"Real Time Detection of Repeated Structures in Point Clouds of Urban Scenes","authors":"Sam Friedman, I. Stamos","doi":"10.1109/3DIMPVT.2011.35","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.35","url":null,"abstract":"Laser range scanners provide rich 3D representations of urban scenes. These scenes are replete with repetitive architectural features such as windows, balconies, and cornices. Processing of dense 3D images is often computationally intensive and occurs offline after acquisition. Here, however, we present an online algorithm for the detection of repetitive features in 3D range scans. Our algorithm creates a function from each scan line by deriving a local measure at each point. Computing the Fourier transform of that function reveals the periodicity of the scene. This robust algorithm forms the basis for novel methods of feature extraction, compression, and registration. What is more this whole process can be executed on-the-fly and integrated into hardware transforming laser scanners into architecture aware devices.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126150668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Finding the Best Feature Detector-Descriptor Combination 寻找最佳特征检测器-描述符组合
A. Dahl, H. Aanæs, K. S. Pedersen
{"title":"Finding the Best Feature Detector-Descriptor Combination","authors":"A. Dahl, H. Aanæs, K. S. Pedersen","doi":"10.1109/3DIMPVT.2011.47","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.47","url":null,"abstract":"Addressing the image correspondence problem by feature matching is a central part of computer vision and 3D inference from images. Consequently, there is a substantial amount of work on evaluating feature detection and feature description methodology. However, the performance of the feature matching is an interplay of both detector and descriptor methodology. Our main contribution is to evaluate the performance of some of the most popular descriptor and detector combinations on the DTU Robot dataset, which is a very large dataset with massive amounts of systematic data aimed at two view matching. The size of the dataset implies that we can also reasonably make deductions about the statistical significance of our results. We conclude, that the MSER and Difference of Gaussian (DoG) detectors with a SIFT or DAISY descriptor are the top performers. This performance is, however, not statistically significantly better than some other methods. As a byproduct of this investigation, we have also tested various DAISY type descriptors, and found that the difference among their performance is statistically insignificant using this dataset. Furthermore, we have not been able to produce results collaborating that using affine invariant feature detectors carries a statistical significant advantage on general scene types.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"275 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127550871","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 57
Hierarchical Shape Matching for Temporally Consistent 3D Video 时间一致性3D视频的分层形状匹配
C. Budd, Peng Huang, A. Hilton
{"title":"Hierarchical Shape Matching for Temporally Consistent 3D Video","authors":"C. Budd, Peng Huang, A. Hilton","doi":"10.1109/3DIMPVT.2011.29","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.29","url":null,"abstract":"In this paper we present a novel approach for temporal alignment of reconstructed mesh sequences with non-rigid surfaces to obtain a consistent representation. We propose a hierarchical scheme for non-sequential matching of frames across the sequence using shape similarity. This gives a tree structure which represents the optimal path for alignment of each frame in the sequence to minimize the change in shape. Non-rigid alignment is performed by recursively traversing the tree to align all frames. Non-sequential alignment reduces problems of drift or tracking failure which occur in previous sequential frame-to-frame techniques. Comparative evaluation on challenging 3D video sequences demonstrates that the proposed approach produces a temporally coherent representation with reduced error in shape and correspondence.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"352 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122849513","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Calibration of Nodal and Free-Moving Cameras in Dynamic Scenes for Post-Production 动态场景中节点和自由移动摄像机的后期校正
E. Imre, Jean-Yves Guillemaut, A. Hilton
{"title":"Calibration of Nodal and Free-Moving Cameras in Dynamic Scenes for Post-Production","authors":"E. Imre, Jean-Yves Guillemaut, A. Hilton","doi":"10.1109/3DIMPVT.2011.40","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.40","url":null,"abstract":"In film production, many post-production tasks require the availability of accurate camera calibration information. This paper presents an algorithm for through-the-lens calibration of a moving camera for a common scenario in film production and broadcasting: The camera views a dynamic scene, which is also viewed by a set of static cameras with known calibration. The proposed method involves the construction of a sparse scene model from the static cameras, with respect to which the moving camera is registered, by applying the appropriate perspective-n-point (PnP) solver. In addition to the general motion case, the algorithm can handle the nodal cameras with unknown focal length via a novel P2P algorithm. The approach can identify a subset of static cameras that are more likely to generate a high number of scene-image correspondences, and can robustly deal with dynamic scenes. Our target applications include dense 3D reconstruction, stereoscopic 3D rendering and 3D scene augmentation, through which the success of the algorithm is demonstrated experimentally.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129654282","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Efficient Generation of Multi-perspective Panoramas 高效的多视角全景生成
Enliang Zheng, R. Raguram, P. Georgel, Jan-Michael Frahm
{"title":"Efficient Generation of Multi-perspective Panoramas","authors":"Enliang Zheng, R. Raguram, P. Georgel, Jan-Michael Frahm","doi":"10.1109/3DIMPVT.2011.60","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.60","url":null,"abstract":"In this paper, we present an efficient technique for generating multi-perspective panoramic images of long scenes. The input to our system is a video sequence captured by a moving camera navigating through a long scene, and our goal is to efficiently generate a panoramic summary of the scene. This problem has received considerable attention in recent years, leading to the development of a number of systems capable of generating high-quality panoramas. However, a significant limitation of current systems is their computational complexity: most current techniques employ computationally expensive algorithms (such as structure-from-motion and dense stereo), or require some degree of manual interaction. In turn, this limits the scalability of the algorithms as well as their ease of implementation. In contrast, the technique we present is simple, efficient, easy to implement, and produces results of comparable quality to state of the art techniques, while doing so at a fraction of the computational cost. Our system operates entirely in the 2D image domain, performing robust image alignment and optical flow based mosaicing, in lieu of more expensive 3D pose/structure computation. We demonstrate the effectiveness of our system on a number of challenging image sequences.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"31 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116338054","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
A System for Capturing Textured 3D Shapes Based on One-Shot Grid Pattern with Multi-band Camera and Infrared Projector 基于单镜头网格模式的多波段相机和红外投影仪纹理三维形状捕捉系统
Kazuhiro Sakashita, Y. Yagi, R. Sagawa, Furukawa Ryo, Hiroshi Kawasaki
{"title":"A System for Capturing Textured 3D Shapes Based on One-Shot Grid Pattern with Multi-band Camera and Infrared Projector","authors":"Kazuhiro Sakashita, Y. Yagi, R. Sagawa, Furukawa Ryo, Hiroshi Kawasaki","doi":"10.1109/3DIMPVT.2011.61","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.61","url":null,"abstract":"Recently, several method have been proposed to capture 3D shapes of a moving object at a high frame rate. One of promising approach to reconstruct a 3D shape is a projector-camera system that projects structured light pattern. One of the problem of this approach is that it has difficulty to obtain texture simultaneously because the texture is interfered by the illumination from the projector. The system proposed in this paper overcomes this issue by separating the light wavelength for texture and shape. The pattern is projected by using infrared light and the texture is captured by using visible light. If the cameras for infrared and visible lights are placed at different position, it causes the misalignment between texture and shape, which degrades the quality of the textured 3D model. Therefore, we developed a multi-band camera that acquires both visible and infrared lights from a single viewpoint. Moreover, to reconstruct a 3D shape using multiple wavelengths of light, namely multiple colors, an infrared pattern projector is developed to generate a multi-band grid pattern. Additionally, a simple method to calibrate the system is proposed by using a fixed grid pattern. Finally, we show the textured 3D shapes captured by the experimental system.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116428246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
An Approach for Shape from Surface Normals with Local Discontinuity Detection 一种基于曲面法线的局部不连续检测方法
Yilin Wang, Enrique Dunn, Jan-Michael Frahm
{"title":"An Approach for Shape from Surface Normals with Local Discontinuity Detection","authors":"Yilin Wang, Enrique Dunn, Jan-Michael Frahm","doi":"10.1109/3DIMPVT.2011.31","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.31","url":null,"abstract":"We present a multi-modal surface reconstruction approach, which utilizes direct surface orientation measurements along with luminance information to obtain high quality 3D reconstructions. The proposed approach models local surface geometry as a set of intersecting natural cubic splines estimated through least squares fitting of our input pixel-wise surface normal measurements. We use this representation to detect discontinuities and segment our scene into disjoint continuous surfaces, which are constructed by an aggregation of connected local surface geometry elements. In order to obtain absolute depth estimates, we introduce the concept of multi-view patch sweeping, where we search for the most photo-consistent patch displacement along a viewing ray. Our approach improves on existing shape from normals methods by enabling absolute depth estimates for scenes with multiple objects. Furthermore, in contrast to existing multi-view stereo methods, we are able to reconstruct textureless regions through the propagation of relative surface orientation measurements. Experiments on synthetic and real data are presented to validate our proposal.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125857507","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3D Reconstruction of Interior Wall Surfaces under Occlusion and Clutter 遮挡和杂波条件下内墙表面的三维重建
A. Adán, Daniel F. Huber
{"title":"3D Reconstruction of Interior Wall Surfaces under Occlusion and Clutter","authors":"A. Adán, Daniel F. Huber","doi":"10.1109/3DIMPVT.2011.42","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.42","url":null,"abstract":"Laser scanners are often used to create 3D models of buildings for civil engineering applications. The current manual process is time-consuming and error-prone. This paper presents a method for using laser scanner data to model predominantly planar surfaces, such as walls, floors, and ceilings, despite the presence of significant amounts of clutter and occlusion, which occur frequently in natural indoor environments. Our goal is to recover the surface shape, detect and model any openings, and fill in the occluded regions. Our method identifies candidate surfaces for modeling, labels occluded surface regions, detects openings in each surface using supervised learning, and reconstructs the surface in the occluded regions. We evaluate the method on a large, highly cluttered data set of a building consisting of forty separate rooms.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"123 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122318260","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 173
An Enhanced 'Optimization-on-a-Manifold' Framework for Global Registration of 3D Range Data 一种增强的“流形优化”框架,用于3D距离数据的全局配准
Francesco Bonarrigo, A. Signoroni
{"title":"An Enhanced 'Optimization-on-a-Manifold' Framework for Global Registration of 3D Range Data","authors":"Francesco Bonarrigo, A. Signoroni","doi":"10.1109/3DIMPVT.2011.51","DOIUrl":"https://doi.org/10.1109/3DIMPVT.2011.51","url":null,"abstract":"In this paper we present a robust global registration technique which is suitable to accurately align sets of high-resolution range images. Our approach is based on the `Optimization-on-a-Manifold', OOM framework proposed by Krishnan et al. to which we contribute with both systemic and computational improvements. The original OOM algorithm performs an error minimization over the manifold of rotations through an iterative scheme based on Gauss-Newton optimization, provided that a set of exact correspondences is known beforehand. As a main contribution, we relax this requirement, allowing to accept sets of inexact correspondences that are dynamically updated after each iteration. Other improvements are directed toward the reduction of the computational burden of the method while maintaining its robustness. The modifications we have introduced allow to significantly improve both the convergence rate and the accuracy of the original technique, while boosting its computational speed. Meaningful comparisons with a classic global registration approach are also provided.","PeriodicalId":330003,"journal":{"name":"2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127460498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信