{"title":"Structured Light Field Design for Correspondence Free Rotation Estimation","authors":"Ian Schillebeeckx, Robert Pless","doi":"10.1109/ICCPHOT.2015.7168376","DOIUrl":"https://doi.org/10.1109/ICCPHOT.2015.7168376","url":null,"abstract":"Many vision and augmented reality applications require knowing the rotation of the camera relative to an object or scene. In this paper we propose to create a structured light field designed explicitly to simplify the estimation of camera rotation. The light field is created using a lenticular sheet with a color coded backplane pattern, creating a light field where the observed color depends on the direction of the light. We show that a picture taken within such a light field gives linear constraints on the K-1R matrix that defines the camera calibration and rotation. In this work we derive an optimization that uses these constraints to rapidly estimate rotation, demonstrate a physical prototype and characterize its sensitivity to errors in the camera focal length and camera color sensitivity.","PeriodicalId":302766,"journal":{"name":"2015 IEEE International Conference on Computational Photography (ICCP)","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124198966","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Directional Super-Resolution by Means of Coded Sampling and Guided Upsampling","authors":"D. Schedl, C. Birklbauer, O. Bimber","doi":"10.1109/ICCPHOT.2015.7168365","DOIUrl":"https://doi.org/10.1109/ICCPHOT.2015.7168365","url":null,"abstract":"We present a simple guided super-resolution technique for increasing directional resolution without reliance on depth estimation or image correspondences. Rather, it searches for best-matching multidimensional (4D or 3D) patches within the entire captured data set to compose new directional images that are consistent in both the spatial and the directional domains. We describe algorithms for guided upsampling, iterative guided upsampling, and sampling code estimation. Our experimental results reveal that the outcomes of existing light-field camera arrays and lightstage systems can be improved without additional hardware requirements or recording effort simply by realignment of cameras or light sources to change their sampling patterns.","PeriodicalId":302766,"journal":{"name":"2015 IEEE International Conference on Computational Photography (ICCP)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132951628","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ronnachai Jaroensri, Sylvain Paris, Aaron Hertzmann, V. Bychkovsky, F. Durand
{"title":"Predicting Range of Acceptable Photographic Tonal Adjustments","authors":"Ronnachai Jaroensri, Sylvain Paris, Aaron Hertzmann, V. Bychkovsky, F. Durand","doi":"10.1109/ICCPHOT.2015.7168372","DOIUrl":"https://doi.org/10.1109/ICCPHOT.2015.7168372","url":null,"abstract":"There is often more than one way to select tonal adjustment-for a photograph, and different individuals may prefer different adjustments. However, selecting good adjustments is challenging. This paper describes a method to predict whether a given tonal rendition is acceptable for a photograph, which we use to characterize its range of acceptable adjustments. We gathered a dataset of image “acceptability” over brightness and contrast adjustments. We find that unacceptable renditions can be explained in terms of overexposure, under-exposure, and low contrast. Based on this observation, we propose a machine-learning algorithm to assess whether an adjusted photograph looks acceptable. We show that our algorithm can differentiate unsightly renditions from reasonable ones. Finally, we describe proof-of-concept applications that use our algorithm to guide the exploration of the possible tonal renditions of a photograph.","PeriodicalId":302766,"journal":{"name":"2015 IEEE International Conference on Computational Photography (ICCP)","volume":"1 6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122585522","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Towards Self-Powered Cameras","authors":"S. Nayar, Daniel C. Sims, M. Fridberg","doi":"10.1109/ICCPHOT.2015.7168377","DOIUrl":"https://doi.org/10.1109/ICCPHOT.2015.7168377","url":null,"abstract":"We propose a simple pixel design, where the pixel's photodiode can be used to not only measure the incident light level, but also to convert the incident light into electrical energy. A sensor architecture is proposed where, during each image capture cycle, the pixels are used first to record and read out the image and then used to harvest energy and charge the sensors' power supply. We have conducted several experiments using off-the-shelf discrete components to validate the practical feasibility of our approach. We first developed a single pixel based on our design and used it to physically scan images of scenes. Next, we developed a fully self-powered camera that produces 30×40 images. The camera uses a supercap rather than an external source as its power supply. For a scene that is around 300 lux in brightness, the voltage across the supercap remains well above the minimum needed for the camera to indefinitely produce an image per second. For scenarios where scene brightness may vary dramatically, we present an adaptive algorithm that adjusts the framerate of the camera based on the voltage of the supercap and the brightness of the scene. Finally, we analyze the light gathering and harvesting properties of our design and explain why we believe it could lead to a fully self-powered solid-state image sensor that produces a useful resolution and framerate.","PeriodicalId":302766,"journal":{"name":"2015 IEEE International Conference on Computational Photography (ICCP)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114677710","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Hang Zhao, Boxin Shi, Christy Fernandez-Cull, Sai-Kit Yeung, R. Raskar
{"title":"Unbounded High Dynamic Range Photography Using a Modulo Camera","authors":"Hang Zhao, Boxin Shi, Christy Fernandez-Cull, Sai-Kit Yeung, R. Raskar","doi":"10.1109/ICCPHOT.2015.7168378","DOIUrl":"https://doi.org/10.1109/ICCPHOT.2015.7168378","url":null,"abstract":"This paper presents a novel framework to extend the dynamic range of images called Unbounded High Dynamic Range (UHDR) photography with a modulo camera. A modulo camera could theoretically take unbounded radiance levels by keeping only the least significant bits. We show that with limited bit depth, very high radiance levels can be recovered from a single modulus image with our newly proposed unwrapping algorithm for natural images. We can also obtain an HDR image with details equally well preserved for all radiance levels by merging the least number of modulus images. Synthetic experiment and experiment with a real modulo camera show the effectiveness of the proposed approach.","PeriodicalId":302766,"journal":{"name":"2015 IEEE International Conference on Computational Photography (ICCP)","volume":"266 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133929960","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Dictionary-Based Approach for Estimating Shape and Spatially-Varying Reflectance","authors":"Zhuo Hui, Aswin C. Sankaranarayanan","doi":"10.1109/ICCPHOT.2015.7168363","DOIUrl":"https://doi.org/10.1109/ICCPHOT.2015.7168363","url":null,"abstract":"We present a technique for estimating the shape and reflectance of an object in terms of its surface normals and spatially-varying BRDF. We assume that multiple images of the object are obtained under fixed view-point and varying illumination, i.e, the setting of photometric stereo. Assuming that the BRDF at each pixel lies in the non-negative span of a known BRDF dictionary, we derive a per-pixel surface normal and BRDF estimation framework that requires neither iterative optimization techniques nor careful initialization, both of which are endemic to most state-of the-art techniques. We showcase the performance of our technique on a wide range of simulated and real scenes where we outperform competing methods.","PeriodicalId":302766,"journal":{"name":"2015 IEEE International Conference on Computational Photography (ICCP)","volume":"177 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123200426","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"LiSens- A Scalable Architecture for Video Compressive Sensing","authors":"Jian Wang, Mohit Gupta, Aswin C. Sankaranarayanan","doi":"10.1109/ICCPHOT.2015.7168369","DOIUrl":"https://doi.org/10.1109/ICCPHOT.2015.7168369","url":null,"abstract":"The measurement rate of cameras that take spatially multiplexed measurements by using spatial light modulators (SLM) is often limited by the switching speed of the SLMs. This is especially true for single-pixel cameras where the photodetector operates at a rate that is many orders-of-magnitude greater than the SLM. We study the factors that determine the measurement rate for such spatial multiplexing cameras (SMC) and show that increasing the number of pixels in the device improves the measurement rate, but there is an optimum number of pixels (typically, few thousands) beyond which the measurement rate does not increase. This motivates the design of LiSens, a novel imaging architecture, that replaces the photodetector in the single-pixel camera with a 1D linear array or a line-sensor. We illustrate the optical architecture underlying LiSens, build a prototype, and demonstrate results of a range of indoor and outdoor scenes. LiSens delivers on the promise of SMCs: imaging at a megapixel resolution, at video rate, using an inexpensive low-resolution sensor.","PeriodicalId":302766,"journal":{"name":"2015 IEEE International Conference on Computational Photography (ICCP)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131333649","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}