M. Becker, M. Baron, D. Kondermann, M. Bußler, V. Helzle
{"title":"Movie dimensionalization via sparse user annotations","authors":"M. Becker, M. Baron, D. Kondermann, M. Bußler, V. Helzle","doi":"10.1109/3DTV.2013.6676633","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676633","url":null,"abstract":"We present a workflow to semi-automatically create depth maps for monocular movie footage. Artists annotate relevant depth discontinuities in a single keyframe. Depth edges are then learned and predicted for the whole shot. We use structure from motion where possible for sparse depth cues, while the artist optionally provides scribbles to improve the intended visual effect. Finally, all three sources of information are combined via variational inpainting scheme. As the outcome of our method is artistic and cannot be evaluated quantitively, we apply our method to a current movie production, showing good results on different scenes. We further evaluate the depth edge localization compared to the “ground truth” provided by artists. To enable experimentation with our approach, we offer our source code.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"2135 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129991337","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"3D mesh and multi-view synthesis implementation using stereo cameras and a depth camera","authors":"Hyok Song, Jisang Yoo, Sooyeong Kwak, Cheon Lee, Byeongho Choi","doi":"10.1109/3DTV.2013.6676645","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676645","url":null,"abstract":"In this paper, we propose a new method of 3D mesh generation and view synthesis for multi-view video generation. The multi-view video can be produced by a view synthesis method and 3D mesh using RGB+Z images which are composed of one or two RGB images and one depth image. We create 36-view images using the view synthesis method and generate 3D mesh with vertexes and faces for free view rendering. Using the image synthesis method, warping, edge compensation, upsampling and 36-view generation processes are used. For 3D mesh, depth and RGB images are converted to the component of 3D mesh data. The quality of synthesized images is measured by MOS(Mean opinion score).","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122637764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
P. Surman, S. Day, K. Akşit, H. Urey, Joshua L. Benjamin, K. Jain, Hao Chen
{"title":"Single and multi-user head tracked glasses-free 3D displays","authors":"P. Surman, S. Day, K. Akşit, H. Urey, Joshua L. Benjamin, K. Jain, Hao Chen","doi":"10.1109/3DTV.2013.6676654","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676654","url":null,"abstract":"This paper describes two head tracked displays; both provide glasses-free 3D (autostereoscopic) to viewers by producing pairs of exit pupils where the left and right stereo images are directed to the appropriate viewers' eyes under the control of a head tracker. The first display is single viewer where exit pupils are formed by a pair of picoprojectors whose position moves in accordance with the viewer's head position. Light is reflected back to the viewer's eyes using a retroreflecting screen. The second display is multi-user and is laser-based. A Gabor superlens screen is scanned with a vertical illumination column that acts as the backlight for a direct-view liquid crystal display (LCD). The emergent beam directions are controlled by a spatial light modulator (SLM) such that they land on the viewers' left and right eyes alternately. For each display the principle of operation, the display hardware and the results are presented here.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116088936","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Amin Banitalebi-Dehkordi, M. Pourazad, P. Nasiopoulos
{"title":"A study on the relationship between depth map quality and the overall 3D video quality of experience","authors":"Amin Banitalebi-Dehkordi, M. Pourazad, P. Nasiopoulos","doi":"10.1109/3DTV.2013.6676650","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676650","url":null,"abstract":"The emergence of multiview displays has made the need for synthesizing virtual views more pronounced, since it is not practical to capture all of the possible views when filming multiview content. View synthesis is performed using the available views and depth maps. There is a correlation between the quality of the synthesized views and the quality of depth maps. In this paper we study the effect of depth map quality on perceptual quality of synthesized view through subjective and objective analysis. Our evaluation results show that: 1) 3D video quality depends highly on the depth map quality and 2) the Visual Information Fidelity index computed between the reference and distorted depth maps has Pearson correlation ratio of 0.75 and Spearman rank order correlation coefficient of 0.67 with the subjective 3D video quality.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129874056","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Computation of virtual environment from stereo-panoramic view","authors":"A. I. Audu, A. Sadka","doi":"10.1109/3DTV.2013.6676659","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676659","url":null,"abstract":"Depth map texturing is one of the prominent image-based rendering methods. It requires the condition of epipolar constraint be satisfied. However, the cylindrical warping involved in the construction of panoramic view makes it highly challenging for this condition to be strictly met. This paper proposes depth-map-based synthesis of virtual environment starting with stereo-panoramic views of a scene. A stereo-panoramic view is first generated by mosaicking a set of images containing both the optical and geometric properties of different parts of the scene. A depth map of the scene contained in the stereo-panoramic view is constructed based on the normalization of matching points and the singular value decomposition of fundamental matrix. Finally, the depth map is textured with one of the panoramic views to obtain a synthesized view. While a good quality depth map can be only generated if and only if epipolar constraint condition for planar rectification is satisfied, it is also the determining factor for the computation of virtual environment from a stereo-panoramic view.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"140 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131690060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"3D conversion using vanishing points and image warping","authors":"Daniel Donatsch, Nico Farber, Matthias Zwicker","doi":"10.1109/3DTV.2013.6676648","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676648","url":null,"abstract":"We describe a user assisted technique for 3D stereo conversion from 2D images. Our approach exploits the geometric structure of perspective images including vanishing points. We allow a user to indicate lines, planes, and vanishing points in the input image, and directly employ these as constraints in an image warping framework to produce a stereo pair. By sidestepping explicit construction of a depth map, our approach is applicable to more general scenes and avoids potential artifacts of depth-image-based rendering. Our method is most suitable for scenes with large scale structures such as buildings.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129538799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A hybrid pixel-block based view synthesis for multiviewpoint 3D video","authors":"Chenxia Li, Ronggang Wang, Wenmin Wang, Ling-yu Duan, Wen Gao","doi":"10.1109/3DTV.2013.6676657","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676657","url":null,"abstract":"View synthesis technique is important for multi-viewpoint 3D video applications. In our technical investigation, we find that some of the current view synthesis methods perform better in boundary areas while some others in non-boundary areas. For example, MPEG VSRS is the most popular view synthesis software, which integrates a series of advanced tools. Recently, Interval-Based Image Synthesis (IBIS) algorithm is proposed, which shows advantage over MPEG VSRS in non-boundary areas but still has some unpleasant artifacts in boundary areas. To take advantage of the merits of the above methods and overcome their shortcomings, we propose a Hybrid Pixel-block Based View Synthesis(HPBVS) method. Experimental results testify that our approach outperforms the state-of-the-art view synthesis methods from both the subjective and objective perspectives.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124021173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
V. K. Adhikarla, A. T. Islam, P. Kovács, O. Staadt
{"title":"Fast and efficient data reduction approach for multi-camera light field display telepresence systems","authors":"V. K. Adhikarla, A. T. Islam, P. Kovács, O. Staadt","doi":"10.1109/3DTV.2013.6676632","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676632","url":null,"abstract":"Cutting-edge telepresence systems equipped with multiple cameras for capturing the whole scene of a collaboration space, face the challenge of transmitting huge amount of dynamic data from multiple viewpoints. With the introduction of Light Field Displays (LFDs) in to the remote collaboration space, it became possible to produce an impression of 3D virtual presence. In addition, LFDs in current generation also rely on the images obtained from cameras arranged in various spatial configurations. To have a realistic and natural 3D collaboration using LFDs, the data in the form of multiple camera images needs to be transmitted in real time using the available bandwidth. Classical compression methods might resolve this issue to a certain level. However, in many cases the achieved compression level is by far insufficient. Moreover, the available compression schemes do not consider any of the display-related attributes. Here, we propose a method by which we reduce the data from each of the camera images by discarding unused parts of the images at the acquisition site in a predetermined way using the display model and geometry, as well as the mapping between the captured and displayed light field. The proposed method is simple to implement and can exclude the unnecessary data in an automatic way. While similar methods exist for 2D screens or display walls, this is the first such algorithm for light fields. Our experimental results show that an identical light field reconstruction can be achieved with the reduced set of data which we would have got if all the data were transmitted. Moreover, the devised method provides very good processing speed.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114256818","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
O. Fatah, A. Aggoun, M. Swash, E. Alazawi, Bei Li, J. C. Fernandez, D. Chen, Emmanuel Tsekleves
{"title":"Generating stereoscopic 3D from holoscopic 3D","authors":"O. Fatah, A. Aggoun, M. Swash, E. Alazawi, Bei Li, J. C. Fernandez, D. Chen, Emmanuel Tsekleves","doi":"10.1109/3DTV.2013.6676638","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676638","url":null,"abstract":"In this paper a single aperture motion picture camera based on holoscopic imaging used to generate high-resolution stereoscopic image. The idea of single aperture camera reduces the very cumbersome and expensive of dual cameras in stereoscopic production. The camera is known as light field camera, which was first proposed in 1908 by lippmann [1]. The rendering method relied on upsampling, shift and integrating of different views to extract stereo images. This is the first experiment attempted to generate stereo form holoscopic content on motion capturing, where researchers so far have been experimenting on still images. In this paper presents motion picture image rendering on holoscopic content to generate content for stereoscopic systems. We have carried out experiments with focused ploneptic camera on a single stage omnidirectional integral camera arrangement with capturing both horizontal and vertical parallax, using a low cost lens array and relay lens. Our results show an improvement in the resolution of images with artifact free and also the rendered stereo content are played back on polarized stereoscopic system and anaglyph system to perceive the 3D depth using filter glasses in our experimental section.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131055398","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}