{"title":"Interactive 2D 3D image conversion method for mobile devices","authors":"H. Emrah Tasli, K. Ugur","doi":"10.1109/3DTV.2011.5877212","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877212","url":null,"abstract":"This work proposes an interactive tool for creating stereo image from a mono image. The user interaction is defined as scribbling on the object of interest followed by relative depth assignment to the selected object. Initial step in the algorithm is to create structured image oversegments with intensity homogeneity and geometrical convexity constraint. The final image segmentation is realized by merging the oversegmented regions in a region growing manner. The proposed local neighbourhood similarity based method saves the energy of searching globally optimum cut. Instead of utilizing an iterative method to search for an optimum energy minimization, local neighbourhood based energy calculation enables one step decision. As the proposed method has relatively low complexity, it could efficiently be used in mobile devices with limited computational resources. The assigned relative depths of multiple objects are used to create stereo image pairs using conventional depth image based rendering (DIBR). The tool has been implemented for Nokia N900 Phone and allows creating stereo images with positive and negative disparity.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126389858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Sjostrom, Peter Hardling, L. Karlsson, R. Olsson
{"title":"Improved depth-image-based rendering algorithm","authors":"M. Sjostrom, Peter Hardling, L. Karlsson, R. Olsson","doi":"10.1109/3DTV.2011.5877183","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877183","url":null,"abstract":"The present work analyses a layered depth-image-based-rendering algorithm based on possible errors occurring with perspective 3D warping. The outcome is improvements to the algorithm that treats depth reliably for scenes containing several levels of foreground objects. The filling of holes of different kinds is addressed so that results have better visual quality. The analysis compares the results of the algorithm with a reference algorithm for the potential error types, and visual examples exhibit the consequences of the improvements. Different objective metrics give ambiguous results, which may be explained by the reduction of structure caused by the reference algorithm.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"636 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131802705","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
K. Vermeirsch, G. Van Hoey, Maurice H.P.H. van Beurden, P. Surman
{"title":"Measurement and evaluation of head tracked auto-stereoscopic displays","authors":"K. Vermeirsch, G. Van Hoey, Maurice H.P.H. van Beurden, P. Surman","doi":"10.1109/3DTV.2011.5877224","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877224","url":null,"abstract":"This paper describes objective and subjective display-related measurement methods to evaluate the performance of a laser illuminated head tracked auto-stereoscopic display. Essential characteristics are speckle, exit pupils and crosstalk. The described methods are used to evaluate a first prototype of an auto-stereoscopic display developed within the European Union-funded HELIUM3D project.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"135 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127568444","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Time-consistency of disocclusion filling algorithms in Depth Image Based Rendering","authors":"Michael Schmeing, Xiaoyi Jiang","doi":"10.1109/3DTV.2011.5877201","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877201","url":null,"abstract":"Depth Image Based Rendering (DIBR) is a technique to render virtual views from the video-plus-depth 3D scene representation. A common phenomenon in DIBR are disocclusions, i.e., unfilled areas that occur in the rendered views. The performance of disocclusion filling algorithms is usually only tested in terms of their ability of finding plausible fillings frame-by-frame. Time-consistency, however, is a crucial property in video processing. Missing time-consistency results in disturbing flicker artifacts. With our work, we want to point out the existence of this problem. We also introduce a measure to quantify the amount of flicker for a first study of this phenomenon.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124058771","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Extended VQM model for predicting 3D video quality considering ambient illumination context","authors":"G. Nur, H. K. Arachchi, S. Dogan, A. Kondoz","doi":"10.1109/3DTV.2011.5877180","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877180","url":null,"abstract":"To speed-up the proliferation of advanced 3-Dimensional (3D) technologies into the consumer market, the influence of these technologies on the perception of 3D video should be determined. Currently, this can only be achieved using either subjective assessment techniques or 2D objective quality evaluation models. Even though the subjective assessment techniques are better than the objective models from the accuracy point of view, they are time consuming and costly. Thus, 2D objective quality evaluation models correlating with Human Visual System (HVS) should be used to predict the 3D video quality perception of users in a reliable way with less effort. Video Quality Metric (VQM), which is a standardized 2D objective quality measurement model due to its well correlation with HVS, is used to predict 3D video quality perception of users reliably. However, ambient illumination context of the viewing environment, which has an effect on 3D video quality perception, is not considered in the quality assessments by VQM. Content adaptation is one of the key applications that need to use the perceived 3D quality assessments under different ambient illumination conditions at regular basis for ensuring improved video experience of users. Therefore, the standardized VQM model is extended using ambient illumination context and content related contexts (i.e., motion, structural feature, and luminance contrast) to predict 3D video quality measurement under a particular ambient illumination condition. The results prove that the extended VQM model can be efficiently utilized to predict the video quality perception of 3D video under a particular ambient illumination condition.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"96 43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129328082","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Viewing window expansion in integral floating display using tilted side mirrors","authors":"J. Yeom, Jae-Hyun Jung, Byoungho Lee","doi":"10.1109/3DTV.2011.5877190","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877190","url":null,"abstract":"We propose an expanding method for viewing window in the integral floating display with the tilted side mirrors. These additional mirrors make some rays change the propagation direction, and the reflected ray has large propagation angle compared with the conventional system. Because the maximum propagation angle of ray determines the viewing window size, the proposed system can enlarge the viewing window successfully. The change of ray density due to the side mirrors is analyzed and the elemental image set generation method using virtual lens array concept is also discussed. The experimental results are provided to verify the performance of the proposed system.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127167895","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Stereo image representation using compressive sensing","authors":"A. S. Akbari, P. B. Zadeh, M. Moniri","doi":"10.1109/3DTV.2011.5877208","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877208","url":null,"abstract":"This paper presents a compressive sensing based stereo image representation technique using wavelet transform gain. The pair of input stereo images is first decomposed into its low-pass and high-pass views using a motion compensated lifting based wavelet transform. A 2D spatial wavelet transform is then further de-correlates the low-pass view into its sub-bands. Wavelet transform gains are employed to regulate threshold value for different sub-bands. The coefficients in high frequency sub-bands and high-pass view are then hard thresholded to generate their sparse sub-bands and view. The compressive sensing method is then used to generate measurements for different resulting sparse sub-bands and view. The baseband coefficients and measurements are finally losslessly coded. The application of compressive sensing in compressing natural images is in its early stages. Therefore, their performances are usually compared with each other than standard codecs. The performance of the proposed codec is superior to the state of the art and is superior to JPEG subjectively.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127238934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Pseudo-curved integral imaging pickup system with realtime view image generation","authors":"G. Park, Mu-Chieh Lo, Byoungho Lee","doi":"10.1109/3DTV.2011.5877211","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877211","url":null,"abstract":"We implement a viewing angle enhanced integral imaging system by the use of more than one lenslet arrays placed in a curve-like manner. The curve effect providing widened viewing angle is achieved, from the nature that we approximate a curved lenslet array using three flat lenslet arrays. To confirm the feasibility of pickup in real time, a digital reconstruction is carried out. It is revealed that the viewing point range of view images is improved considerably compared with that of the conventional integral imaging system.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"434 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116060578","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Asymmetric spatial scalability in stereoscopic video coding","authors":"Junjian Quan, M. Hannuksela, Houqiang Li","doi":"10.1109/3DTV.2011.5877219","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877219","url":null,"abstract":"Under many viewing circumstances, asymmetric stereoscopic video coding provides similar perceived quality with significant decrease in bitrate, computational complexity, and memory usage compared to conventional stereoscopic video coding. However, in other circumstances, users may prefer conventional symmetric stereoscopic video to asymmetric stereoscopic video. This paper presents a stereoscopic video coding scheme including asymmetric spatial scalability to provide the bitstream adaptation capability to meet heterogeneous network conditions, receiver devices, and user preferences. The spatial scalability property is enabled for the non-base view, and hence the coded bitstream can be decoded as conventional or mixed-resolution stereoscopic video. Simulation results show that the proposed method outperforms the tested other coding schemes providing the same scalability property by at least 8.2% in terms of average bitrate saving.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121529704","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Fengli Yu, Ju Liu, Yannan Ren, Jiande Sun, Yuling Gao, Wei Liu
{"title":"Depth generation method for 2D to 3D conversion","authors":"Fengli Yu, Ju Liu, Yannan Ren, Jiande Sun, Yuling Gao, Wei Liu","doi":"10.1109/3DTV.2011.5877196","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877196","url":null,"abstract":"An efficient depth map generation method is presented for static scenes with moving objects. Firstly, static background scene is reconstructed. Depth map of the reconstructed static background scene is extracted by linear perspective. Then, moving objects are segmented precisely. Depth values are assigned to the segmented moving objects according to their positions in the static scene. Finally, the depth values of the static background scene and the moving objects are integrated into one depth map. Experimental results show that the proposed method can generate smooth and reliable depth maps.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127798507","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}