CVPR 2011Pub Date : 2011-06-20DOI: 10.1109/CVPR.2011.5995440
Jean-Sébastien Franco, Edmond Boyer
{"title":"Learning temporally consistent rigidities","authors":"Jean-Sébastien Franco, Edmond Boyer","doi":"10.1109/CVPR.2011.5995440","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995440","url":null,"abstract":"We present a novel probabilistic framework for rigid tracking and segmentation of shapes observed from multiple cameras. Most existing methods have focused on solving each of these problems individually, segmenting the shape assuming surface registration is solved, or conversely performing surface registration assuming shape segmentation or kinematic structure is known. We assume no prior kinematic or registration knowledge except for an over-estimate k of the number of rigidities in the scene, instead proposing to simultaneously discover, adapt, and track its rigid structure on the fly. We simultaneously segment and infer poses of rigid subcomponents of a single chosen reference mesh acquired in the sequence. We show that this problem can be rigorously cast as a likelihood maximization over rigid component parameters. We solve this problem using an Expectation Maximization algorithm, with latent observation assignments to reference vertices and rigid parts. Our experiments on synthetic and real data show the validity of the method, robustness to noise, and its promising applicability to complex sequences.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123563540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
CVPR 2011Pub Date : 2011-06-20DOI: 10.1109/CVPR.2011.5995549
Richard Roberts, Sudipta N. Sinha, R. Szeliski, Drew Steedly
{"title":"Structure from motion for scenes with large duplicate structures","authors":"Richard Roberts, Sudipta N. Sinha, R. Szeliski, Drew Steedly","doi":"10.1109/CVPR.2011.5995549","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995549","url":null,"abstract":"Most existing structure from motion (SFM) approaches for unordered images cannot handle multiple instances of the same structure in the scene. When image pairs containing different instances are matched based on visual similarity, the pairwise geometric relations as well as the correspondences inferred from such pairs are erroneous, which can lead to catastrophic failures in the reconstruction. In this paper, we investigate the geometric ambiguities caused by the presence of repeated or duplicate structures and show that to disambiguate between multiple hypotheses requires more than pure geometric reasoning. We couple an expectation maximization (EM)-based algorithm that estimates camera poses and identifies the false match-pairs with an efficient sampling method to discover plausible data association hypotheses. The sampling method is informed by geometric and image-based cues. Our algorithm usually recovers the correct data association, even in the presence of large numbers of false pairwise matches.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122017565","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
CVPR 2011Pub Date : 2011-06-20DOI: 10.1109/CVPR.2011.5995703
M. Dixon, Austin Abrams, Nathan Jacobs, Robert Pless
{"title":"On analyzing video with very small motions","authors":"M. Dixon, Austin Abrams, Nathan Jacobs, Robert Pless","doi":"10.1109/CVPR.2011.5995703","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995703","url":null,"abstract":"We characterize a class of videos consisting of very small but potentially complicated motions. We find that in these scenes, linear appearance variations have a direct relationship to scene motions. We show how to interpret appearance variations captured through a PCA decomposition of the image set as a scene-specific non-parametric motion basis. We propose fast, robust tools for dense flow estimates that are effective in scenes with small motions and potentially large image noise. We show example results in a variety of applications, including motion segmentation and long-term point tracking.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"50 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116822933","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
CVPR 2011Pub Date : 2011-06-20DOI: 10.1109/CVPR.2011.5995473
Xiaowu Chen, Mengmeng Chen, Xin Jin, Qinping Zhao
{"title":"Face illumination transfer through edge-preserving filters","authors":"Xiaowu Chen, Mengmeng Chen, Xin Jin, Qinping Zhao","doi":"10.1109/CVPR.2011.5995473","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995473","url":null,"abstract":"This article proposes a novel image-based method to transfer illumination from a reference face image to a target face image through edge-preserving filters. According to our method, only a single reference image, without any knowledge of the 3D geometry or material information of the target face, is needed. We first decompose the lightness layers of the reference and the target images into large-scale and detail layers through weighted least square (WLS) filter after face alignment. The large-scale layer of the reference image is filtered with the guidance of the target image. Adaptive parameter selection schemes for the edge-preserving filters is proposed in the above two filtering steps. The final relit result is obtained by replacing the large-scale layer of the target image with that of the reference image. We acquire convincing relit result on numerous target and reference face images with different lighting effects and genders. Comparisons with previous work show that our method is less affected by geometry differences and can preserve better the identification structure and skin color of the target face.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124029381","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
CVPR 2011Pub Date : 2011-06-20DOI: 10.1109/CVPR.2011.5995518
Junfeng He, Shih-Fu Chang, R. Radhakrishnan, C. Bauer
{"title":"Compact hashing with joint optimization of search accuracy and time","authors":"Junfeng He, Shih-Fu Chang, R. Radhakrishnan, C. Bauer","doi":"10.1109/CVPR.2011.5995518","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995518","url":null,"abstract":"Similarity search, namely, finding approximate nearest neighborhoods, is the core of many large scale machine learning or vision applications. Recently, many research results demonstrate that hashing with compact codes can achieve promising performance for large scale similarity search. However, most of the previous hashing methods with compact codes only model and optimize the search accuracy. Search time, which is an important factor for hashing in practice, is usually not addressed explicitly. In this paper, we develop a new scalable hashing algorithm with joint optimization of search accuracy and search time simultaneously. Our method generates compact hash codes for data of general formats with any similarity function. We evaluate our method using diverse data sets up to 1 million samples (e.g., web images). Our comprehensive results show the proposed method significantly outperforms several state-of-the-art hashing approaches.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"166 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125969451","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
CVPR 2011Pub Date : 2011-06-20DOI: 10.1109/CVPR.2011.5995579
Ilya Reshetouski, A. Manakov, H. Seidel, Ivo Ihrke
{"title":"Three-dimensional kaleidoscopic imaging","authors":"Ilya Reshetouski, A. Manakov, H. Seidel, Ivo Ihrke","doi":"10.1109/CVPR.2011.5995579","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995579","url":null,"abstract":"We introduce three-dimensional kaleidoscopic imaging, a promising alternative for recording multi-view imagery. The main limitation of multi-view reconstruction techniques is the limited number of views that are available from multi-camera systems, especially for dynamic scenes. Our new system is based on imaging an object inside a kaleidoscopic mirror system. We show that this approach can generate a large number of high-quality views well distributed over the hemisphere surrounding the object in a single shot. In comparison to existing multi-view systems, our method offers a number of advantages: it is possible to operate with a single camera, the individual views are perfectly synchronized, and they have the same radiometric and colorimetric properties. We describe the setup both theoretically, and provide methods for a practical implementation. Enabling interfacing to standard multi-view algorithms for further processing is an important goal of our techniques.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124693362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
CVPR 2011Pub Date : 2011-06-20DOI: 10.1109/CVPR.2011.5995388
Chenglei Wu, Bennett Wilburn, Yasuyuki Matsushita, Christian Theobalt
{"title":"High-quality shape from multi-view stereo and shading under general illumination","authors":"Chenglei Wu, Bennett Wilburn, Yasuyuki Matsushita, Christian Theobalt","doi":"10.1109/CVPR.2011.5995388","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995388","url":null,"abstract":"Multi-view stereo methods reconstruct 3D geometry from images well for sufficiently textured scenes, but often fail to recover high-frequency surface detail, particularly for smoothly shaded surfaces. On the other hand, shape-from-shading methods can recover fine detail from shading variations. Unfortunately, it is non-trivial to apply shape-from-shading alone to multi-view data, and most shading-based estimation methods only succeed under very restricted or controlled illumination. We present a new algorithm that combines multi-view stereo and shading-based refinement for high-quality reconstruction of 3D geometry models from images taken under constant but otherwise arbitrary illumination. We have tested our algorithm on several scenes that were captured under several general and unknown lighting conditions, and we show that our final reconstructions rival laser range scans.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129384016","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
CVPR 2011Pub Date : 2011-06-20DOI: 10.1109/CVPR.2011.5995321
Mohit Gupta, Amit K. Agrawal, A. Veeraraghavan, S. Narasimhan
{"title":"Structured light 3D scanning in the presence of global illumination","authors":"Mohit Gupta, Amit K. Agrawal, A. Veeraraghavan, S. Narasimhan","doi":"10.1109/CVPR.2011.5995321","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995321","url":null,"abstract":"Global illumination effects such as inter-reflections, diffusion and sub-surface scattering severely degrade the performance of structured light-based 3D scanning. In this paper, we analyze the errors caused by global illumination in structured light-based shape recovery. Based on this analysis, we design structured light patterns that are resilient to individual global illumination effects using simple logical operations and tools from combinatorial mathematics. Scenes exhibiting multiple phenomena are handled by combining results from a small ensemble of such patterns. This combination also allows us to detect any residual errors that are corrected by acquiring a few additional images. Our techniques do not require explicit separation of the direct and global components of scene radiance and hence work even in scenarios where the separation fails or the direct component is too low. Our methods can be readily incorporated into existing scanning systems without significant overhead in terms of capture time or hardware. We show results on a variety of scenes with complex shape and material properties and challenging global illumination effects.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129724231","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
CVPR 2011Pub Date : 2011-06-20DOI: 10.1109/CVPR.2011.5995533
M. Hansard, R. Horaud, Michel Amat, Seungkyu Lee
{"title":"Projective alignment of range and parallax data","authors":"M. Hansard, R. Horaud, Michel Amat, Seungkyu Lee","doi":"10.1109/CVPR.2011.5995533","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995533","url":null,"abstract":"An approximately Euclidean representation of the visible scene can be obtained directly from a range, or ‘time-of-flight’, camera. An uncalibrated binocular system, in contrast, gives only a projective reconstruction of the scene. This paper analyzes the geometric mapping between the two representations, without requiring an intermediate calibration of the binocular system. The mapping can be found by either of two new methods, one of which requires point-correspondences between the range and colour cameras, and one of which does not. It is shown that these methods can be used to reproject the range data into the binocular images, which makes it possible to associate high-resolution colour and texture with each point in the Euclidean representation.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128255933","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
CVPR 2011Pub Date : 2011-06-20DOI: 10.1109/CVPR.2011.5995418
Xun Cao, Xin Tong, Qionghai Dai, Stephen Lin
{"title":"High resolution multispectral video capture with a hybrid camera system","authors":"Xun Cao, Xin Tong, Qionghai Dai, Stephen Lin","doi":"10.1109/CVPR.2011.5995418","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995418","url":null,"abstract":"We present a new approach to capture video at high spatial and spectral resolutions using a hybrid camera system. Composed of an RGB video camera, a grayscale video camera and several optical elements, the hybrid camera system simultaneously records two video streams: an RGB video with high spatial resolution, and a multispectral video with low spatial resolution. After registration of the two video streams, our system propagates the multispectral information into the RGB video to produce a video with both high spectral and spatial resolution. This propagation between videos is guided by color similarity of pixels in the spectral domain, proximity in the spatial domain, and the consistent color of each scene point in the temporal domain. The propagation algorithm is designed for rapid computation to allow real-time video generation at the original frame rate, and can thus facilitate real-time video analysis tasks such as tracking and surveillance. Hardware implementation details and design tradeoffs are discussed. We evaluate the proposed system using both simulations with ground truth data and on real-world scenes. The utility of this high resolution multispectral video data is demonstrated in dynamic white balance adjustment and tracking.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128273468","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}