2013 3DTV Vision Beyond Depth (3DTV-CON)最新文献

筛选
英文 中文
Movie dimensionalization via sparse user annotations 通过稀疏用户注释实现电影维度化
2013 3DTV Vision Beyond Depth (3DTV-CON) Pub Date : 2013-12-02 DOI: 10.1109/3DTV.2013.6676633
M. Becker, M. Baron, D. Kondermann, M. Bußler, V. Helzle
{"title":"Movie dimensionalization via sparse user annotations","authors":"M. Becker, M. Baron, D. Kondermann, M. Bußler, V. Helzle","doi":"10.1109/3DTV.2013.6676633","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676633","url":null,"abstract":"We present a workflow to semi-automatically create depth maps for monocular movie footage. Artists annotate relevant depth discontinuities in a single keyframe. Depth edges are then learned and predicted for the whole shot. We use structure from motion where possible for sparse depth cues, while the artist optionally provides scribbles to improve the intended visual effect. Finally, all three sources of information are combined via variational inpainting scheme. As the outcome of our method is artistic and cannot be evaluated quantitively, we apply our method to a current movie production, showing good results on different scenes. We further evaluate the depth edge localization compared to the “ground truth” provided by artists. To enable experimentation with our approach, we offer our source code.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"2135 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129991337","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
3D mesh and multi-view synthesis implementation using stereo cameras and a depth camera 使用立体相机和深度相机实现3D网格和多视图合成
2013 3DTV Vision Beyond Depth (3DTV-CON) Pub Date : 2013-12-02 DOI: 10.1109/3DTV.2013.6676645
Hyok Song, Jisang Yoo, Sooyeong Kwak, Cheon Lee, Byeongho Choi
{"title":"3D mesh and multi-view synthesis implementation using stereo cameras and a depth camera","authors":"Hyok Song, Jisang Yoo, Sooyeong Kwak, Cheon Lee, Byeongho Choi","doi":"10.1109/3DTV.2013.6676645","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676645","url":null,"abstract":"In this paper, we propose a new method of 3D mesh generation and view synthesis for multi-view video generation. The multi-view video can be produced by a view synthesis method and 3D mesh using RGB+Z images which are composed of one or two RGB images and one depth image. We create 36-view images using the view synthesis method and generate 3D mesh with vertexes and faces for free view rendering. Using the image synthesis method, warping, edge compensation, upsampling and 36-view generation processes are used. For 3D mesh, depth and RGB images are converted to the component of 3D mesh data. The quality of synthesized images is measured by MOS(Mean opinion score).","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122637764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Single and multi-user head tracked glasses-free 3D displays 单用户和多用户头跟踪无眼镜3D显示
2013 3DTV Vision Beyond Depth (3DTV-CON) Pub Date : 2013-12-02 DOI: 10.1109/3DTV.2013.6676654
P. Surman, S. Day, K. Akşit, H. Urey, Joshua L. Benjamin, K. Jain, Hao Chen
{"title":"Single and multi-user head tracked glasses-free 3D displays","authors":"P. Surman, S. Day, K. Akşit, H. Urey, Joshua L. Benjamin, K. Jain, Hao Chen","doi":"10.1109/3DTV.2013.6676654","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676654","url":null,"abstract":"This paper describes two head tracked displays; both provide glasses-free 3D (autostereoscopic) to viewers by producing pairs of exit pupils where the left and right stereo images are directed to the appropriate viewers' eyes under the control of a head tracker. The first display is single viewer where exit pupils are formed by a pair of picoprojectors whose position moves in accordance with the viewer's head position. Light is reflected back to the viewer's eyes using a retroreflecting screen. The second display is multi-user and is laser-based. A Gabor superlens screen is scanned with a vertical illumination column that acts as the backlight for a direct-view liquid crystal display (LCD). The emergent beam directions are controlled by a spatial light modulator (SLM) such that they land on the viewers' left and right eyes alternately. For each display the principle of operation, the display hardware and the results are presented here.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116088936","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A study on the relationship between depth map quality and the overall 3D video quality of experience 深度图质量与整体3D视频体验质量的关系研究
2013 3DTV Vision Beyond Depth (3DTV-CON) Pub Date : 2013-12-02 DOI: 10.1109/3DTV.2013.6676650
Amin Banitalebi-Dehkordi, M. Pourazad, P. Nasiopoulos
{"title":"A study on the relationship between depth map quality and the overall 3D video quality of experience","authors":"Amin Banitalebi-Dehkordi, M. Pourazad, P. Nasiopoulos","doi":"10.1109/3DTV.2013.6676650","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676650","url":null,"abstract":"The emergence of multiview displays has made the need for synthesizing virtual views more pronounced, since it is not practical to capture all of the possible views when filming multiview content. View synthesis is performed using the available views and depth maps. There is a correlation between the quality of the synthesized views and the quality of depth maps. In this paper we study the effect of depth map quality on perceptual quality of synthesized view through subjective and objective analysis. Our evaluation results show that: 1) 3D video quality depends highly on the depth map quality and 2) the Visual Information Fidelity index computed between the reference and distorted depth maps has Pearson correlation ratio of 0.75 and Spearman rank order correlation coefficient of 0.67 with the subjective 3D video quality.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129874056","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Computation of virtual environment from stereo-panoramic view 基于立体全景的虚拟环境计算
2013 3DTV Vision Beyond Depth (3DTV-CON) Pub Date : 2013-12-02 DOI: 10.1109/3DTV.2013.6676659
A. I. Audu, A. Sadka
{"title":"Computation of virtual environment from stereo-panoramic view","authors":"A. I. Audu, A. Sadka","doi":"10.1109/3DTV.2013.6676659","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676659","url":null,"abstract":"Depth map texturing is one of the prominent image-based rendering methods. It requires the condition of epipolar constraint be satisfied. However, the cylindrical warping involved in the construction of panoramic view makes it highly challenging for this condition to be strictly met. This paper proposes depth-map-based synthesis of virtual environment starting with stereo-panoramic views of a scene. A stereo-panoramic view is first generated by mosaicking a set of images containing both the optical and geometric properties of different parts of the scene. A depth map of the scene contained in the stereo-panoramic view is constructed based on the normalization of matching points and the singular value decomposition of fundamental matrix. Finally, the depth map is textured with one of the panoramic views to obtain a synthesized view. While a good quality depth map can be only generated if and only if epipolar constraint condition for planar rectification is satisfied, it is also the determining factor for the computation of virtual environment from a stereo-panoramic view.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"140 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131690060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3D conversion using vanishing points and image warping 使用消失点和图像扭曲的3D转换
2013 3DTV Vision Beyond Depth (3DTV-CON) Pub Date : 2013-10-07 DOI: 10.1109/3DTV.2013.6676648
Daniel Donatsch, Nico Farber, Matthias Zwicker
{"title":"3D conversion using vanishing points and image warping","authors":"Daniel Donatsch, Nico Farber, Matthias Zwicker","doi":"10.1109/3DTV.2013.6676648","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676648","url":null,"abstract":"We describe a user assisted technique for 3D stereo conversion from 2D images. Our approach exploits the geometric structure of perspective images including vanishing points. We allow a user to indicate lines, planes, and vanishing points in the input image, and directly employ these as constraints in an image warping framework to produce a stereo pair. By sidestepping explicit construction of a depth map, our approach is applicable to more general scenes and avoids potential artifacts of depth-image-based rendering. Our method is most suitable for scenes with large scale structures such as buildings.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129538799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A hybrid pixel-block based view synthesis for multiviewpoint 3D video 基于混合像素块的多视点3D视频视图合成
2013 3DTV Vision Beyond Depth (3DTV-CON) Pub Date : 2013-10-01 DOI: 10.1109/3DTV.2013.6676657
Chenxia Li, Ronggang Wang, Wenmin Wang, Ling-yu Duan, Wen Gao
{"title":"A hybrid pixel-block based view synthesis for multiviewpoint 3D video","authors":"Chenxia Li, Ronggang Wang, Wenmin Wang, Ling-yu Duan, Wen Gao","doi":"10.1109/3DTV.2013.6676657","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676657","url":null,"abstract":"View synthesis technique is important for multi-viewpoint 3D video applications. In our technical investigation, we find that some of the current view synthesis methods perform better in boundary areas while some others in non-boundary areas. For example, MPEG VSRS is the most popular view synthesis software, which integrates a series of advanced tools. Recently, Interval-Based Image Synthesis (IBIS) algorithm is proposed, which shows advantage over MPEG VSRS in non-boundary areas but still has some unpleasant artifacts in boundary areas. To take advantage of the merits of the above methods and overcome their shortcomings, we propose a Hybrid Pixel-block Based View Synthesis(HPBVS) method. Experimental results testify that our approach outperforms the state-of-the-art view synthesis methods from both the subjective and objective perspectives.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124021173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Fast and efficient data reduction approach for multi-camera light field display telepresence systems 多摄像头光场显示远程临场感系统的快速高效数据化简方法
2013 3DTV Vision Beyond Depth (3DTV-CON) Pub Date : 2013-10-01 DOI: 10.1109/3DTV.2013.6676632
V. K. Adhikarla, A. T. Islam, P. Kovács, O. Staadt
{"title":"Fast and efficient data reduction approach for multi-camera light field display telepresence systems","authors":"V. K. Adhikarla, A. T. Islam, P. Kovács, O. Staadt","doi":"10.1109/3DTV.2013.6676632","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676632","url":null,"abstract":"Cutting-edge telepresence systems equipped with multiple cameras for capturing the whole scene of a collaboration space, face the challenge of transmitting huge amount of dynamic data from multiple viewpoints. With the introduction of Light Field Displays (LFDs) in to the remote collaboration space, it became possible to produce an impression of 3D virtual presence. In addition, LFDs in current generation also rely on the images obtained from cameras arranged in various spatial configurations. To have a realistic and natural 3D collaboration using LFDs, the data in the form of multiple camera images needs to be transmitted in real time using the available bandwidth. Classical compression methods might resolve this issue to a certain level. However, in many cases the achieved compression level is by far insufficient. Moreover, the available compression schemes do not consider any of the display-related attributes. Here, we propose a method by which we reduce the data from each of the camera images by discarding unused parts of the images at the acquisition site in a predetermined way using the display model and geometry, as well as the mapping between the captured and displayed light field. The proposed method is simple to implement and can exclude the unnecessary data in an automatic way. While similar methods exist for 2D screens or display walls, this is the first such algorithm for light fields. Our experimental results show that an identical light field reconstruction can be achieved with the reduced set of data which we would have got if all the data were transmitted. Moreover, the devised method provides very good processing speed.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114256818","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Generating stereoscopic 3D from holoscopic 3D 从全息3D生成立体3D
2013 3DTV Vision Beyond Depth (3DTV-CON) Pub Date : 2013-10-01 DOI: 10.1109/3DTV.2013.6676638
O. Fatah, A. Aggoun, M. Swash, E. Alazawi, Bei Li, J. C. Fernandez, D. Chen, Emmanuel Tsekleves
{"title":"Generating stereoscopic 3D from holoscopic 3D","authors":"O. Fatah, A. Aggoun, M. Swash, E. Alazawi, Bei Li, J. C. Fernandez, D. Chen, Emmanuel Tsekleves","doi":"10.1109/3DTV.2013.6676638","DOIUrl":"https://doi.org/10.1109/3DTV.2013.6676638","url":null,"abstract":"In this paper a single aperture motion picture camera based on holoscopic imaging used to generate high-resolution stereoscopic image. The idea of single aperture camera reduces the very cumbersome and expensive of dual cameras in stereoscopic production. The camera is known as light field camera, which was first proposed in 1908 by lippmann [1]. The rendering method relied on upsampling, shift and integrating of different views to extract stereo images. This is the first experiment attempted to generate stereo form holoscopic content on motion capturing, where researchers so far have been experimenting on still images. In this paper presents motion picture image rendering on holoscopic content to generate content for stereoscopic systems. We have carried out experiments with focused ploneptic camera on a single stage omnidirectional integral camera arrangement with capturing both horizontal and vertical parallax, using a low cost lens array and relay lens. Our results show an improvement in the resolution of images with artifact free and also the rendered stereo content are played back on polarized stereoscopic system and anaglyph system to perceive the 3D depth using filter glasses in our experimental section.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131055398","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信