2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)最新文献

筛选
英文 中文
Immersive 3D user interface for 3D TVS 3D电视的沉浸式3D用户界面
Lin Du, Peng Qin, Jianping Song, Wenjuan Song, Yan Xu, Wei Zhou
{"title":"Immersive 3D user interface for 3D TVS","authors":"Lin Du, Peng Qin, Jianping Song, Wenjuan Song, Yan Xu, Wei Zhou","doi":"10.1109/3DTV.2011.5877182","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877182","url":null,"abstract":"This paper investigates some important visual factors of immersive 3D user interfaces on the latest 3D displays and studies how they can be applied for 3D TVs at home. The vergence-accommodation conflict in immersive 3D user interfaces is addressed in order to theoretically estimate some perceived parameters of the user interfaces under this constraint. To demonstrate the proposed immersive 3D user interfaces, a 3D TV browse and control demo is built and some stereo screenshot results are presented.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116555971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A fast image segmentation algorithm using color and depth map 使用颜色和深度图的快速图像分割算法
Emanuele Mirante, M. Georgiev, A. Gotchev
{"title":"A fast image segmentation algorithm using color and depth map","authors":"Emanuele Mirante, M. Georgiev, A. Gotchev","doi":"10.1109/3DTV.2011.5877227","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877227","url":null,"abstract":"In this paper, a real-time image segmentation algorithm is presented. It utilizes both color and depth information retrieved from a multi-sensor capture system, which combines stereo camera pairs with time-of-flight range sensor. The algorithm targets low complexity and fast implementation which can be achieved through parallelization. Applications, such as immersive videoconferencing and lecturer segmentation for augmented reality lecture presentation, can benefit from the designed algorithm.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"122 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115961749","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Depth map generation for 2D-to-3D conversion by limited user inputs and depth propagation 通过有限的用户输入和深度传播生成2d到3d转换的深度图
Xi Yan, You Yang, Guihua Er, Qionghai Dai
{"title":"Depth map generation for 2D-to-3D conversion by limited user inputs and depth propagation","authors":"Xi Yan, You Yang, Guihua Er, Qionghai Dai","doi":"10.1109/3DTV.2011.5877167","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877167","url":null,"abstract":"The quality of depth maps is crucial for 2D to 3D conversion, but high quality depth map generation methods are usually very time consuming. We propose an efficient semi-automatic depth map generation scheme based on limited user inputs and depth propagation. First, the original image is over-segmented. Then, the depth values of selected pixels and the approximate locations of T-junctions are specified by user inputs. The final depth map is obtained by depth propagation combining user inputs, color and edge information. The experimental results demonstrate that our scheme is satisfactory in terms of both accuracy and efficiency, and thus can be applied for high quality 2D to 3D video conversion.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125953540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Robust focal length estimation based on minimal solution method 基于最小解法的鲁棒焦距估计
Deqing Chen, Hang Shao, Qionghai Dai
{"title":"Robust focal length estimation based on minimal solution method","authors":"Deqing Chen, Hang Shao, Qionghai Dai","doi":"10.1109/3DTV.2011.5877179","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877179","url":null,"abstract":"We present a new approach to estimate the focal length for camera calibration in multiview reconstruction. As a popular camera calibration approach, minimal solution method gives rise to a great number of focal-length estimates, from which generating an accurate one is of great significance. Our method concentrates on how to obtain an accurate estimate and is carried out in two steps: firstly, a norm constraint for the fundamental matrix is employed to prune the low-confidence focal-length candidates. Then the focal-length estimate is obtained with a robust focal-length estimation scheme, which consists of occurrence to probability transform, focal-length candidates resample and final estimation with expectation. Experimental results demonstrate that our method could obtain better estimate with both higher accuracy and higher stability than the state-of-the-art method.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134141730","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Self-adaptive normal estimation and position adjustment for MVS reconstruction MVS重建的自适应正态估计和位置调整
Yanjun Qian, Qionghai Dai, Guihua Er
{"title":"Self-adaptive normal estimation and position adjustment for MVS reconstruction","authors":"Yanjun Qian, Qionghai Dai, Guihua Er","doi":"10.1109/3DTV.2011.5877161","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877161","url":null,"abstract":"Generating a polygonal mesh model from the point cloud is a critical step of many state-of-art MVS reconstruction algorithms, and influences the accuracy and visual quality of the final results significantly. The normal estimation and position adjustment of each point is required for this procedure. We present a mathematical analysis of the normal estimation approach, and propose two hypotheses to determinate the accuracy and smoothness of the points in a local region. A multi-scale strategy is implemented to obtain a proper scale for each point. Then the according normal is calculated by PCA on this scale, and the positions can be optimized by combining the accurate neighboring normals. A 2D toy example proves that the proposed approach can adjust the noisy point to the right surface while preserving details. At last we show that our method can improve the quality of mesh models for real MVS reconstruction tasks.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131877861","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimation of subjective quality for mixed-resolution stereoscopic video 混合分辨率立体视频主观质量的估计
P. Aflaki, M. Hannuksela, Jussi H. Hakala, J. Hakkinen, M. Gabbouj
{"title":"Estimation of subjective quality for mixed-resolution stereoscopic video","authors":"P. Aflaki, M. Hannuksela, Jussi H. Hakala, J. Hakkinen, M. Gabbouj","doi":"10.1109/3DTV.2011.5877171","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877171","url":null,"abstract":"In mixed-resolution (MR) stereoscopic video, one view is presented with a lower resolution compared with the other one; therefore, a lower bitrate, a reduced computational complexity, and a decrease in memory access bandwidth can be expected in coding. The human visual system is known to fuse left and right views in such a way that the perceptual visual quality is closer to that of the higher-resolution view. In this paper, a subjective assessment of mixed resolution (MR) stereoscopic videos is presented and the results are analyzed and compared with previous subjective tests presented in the literature. Three downsampling ratios 1/2, 3/8, and 1/4 were used to create lower-resolution views. Hence, the lower-resolution view had different spatial resolutions in terms of pixels per degree (PPD) for each downsampling ratio. It was discovered that the subjective viewing experience tended to follow a logarithmic function of the spatial resolution of the lower-resolution view measured in PPD. A similar behavior was also found from the results of an earlier experiment. Thus, the results suggest that the presented logarithmic function characterizes the expected viewing experience of MR stereoscopic video.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132829060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Evaluation of perceived depth resolution in multi-view threedimensional display using depth image-based rendering 基于深度图像渲染的多视图三维显示感知深度分辨率评估
Jae-Hyun Jung, Jisoo Hong, G. Park, Keehoon Hong, S. Min, Byoungho Lee
{"title":"Evaluation of perceived depth resolution in multi-view threedimensional display using depth image-based rendering","authors":"Jae-Hyun Jung, Jisoo Hong, G. Park, Keehoon Hong, S. Min, Byoungho Lee","doi":"10.1109/3DTV.2011.5877156","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877156","url":null,"abstract":"This paper describes the threshold of perception of depth resolution in multi-view three-dimensional display using depth image-based rendering (DIBR). In the pickup process, we capture the stereo image and center depth map in various depth resolutions using OpenGL. The multi-view images are generated by the depth image-based rendering of the stereo image and center depth map in different depth resolution. The synthesized view-images and simulation results are compared with regard to peak signal-to-noise ratio (PSNR), and the perception threshold of depth resolution is defined by the result of questionnaire with stereoscopic display.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130608328","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Depth up-sampling for depth coding using view information 使用视图信息进行深度编码的深度上采样
M. O. Wildeboer, T. Yendo, M. Panahpour Tehrani, T. Fujii, M. Tanimoto
{"title":"Depth up-sampling for depth coding using view information","authors":"M. O. Wildeboer, T. Yendo, M. Panahpour Tehrani, T. Fujii, M. Tanimoto","doi":"10.1109/3DTV.2011.5877198","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877198","url":null,"abstract":"In 3DTV and free viewpoint imaging systems based on a view-plus-depth representation, depth compression is important for high-quality view synthesis. Several publications have proposed depth down-/up-sampling as part of the depth coding strategy. Recently, we proposed a depth up-sampling method which uses the high resolution view in the process of depth up-sampling. Actually, in 2007 Kopf et.al. already proposed a Joint Bilateral Upsampler(JBU), which uses a high resolution input image as prior in a joint bilateral filter for upscaling. In this paper we compare our previous method with the JBU approach, in the context of depth coding. Furthermore, we analyze the influence of view coding on the depth up-sampling result.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115205520","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Decoder-side macroblock information derivation for efficient multiview video plus depth map coding 解码器侧宏块信息派生,用于高效的多视点视频和深度图编码
S. Shimizu, H. Kimata, Shiori Sugimoto, N. Matsuura
{"title":"Decoder-side macroblock information derivation for efficient multiview video plus depth map coding","authors":"S. Shimizu, H. Kimata, Shiori Sugimoto, N. Matsuura","doi":"10.1109/3DTV.2011.5877205","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877205","url":null,"abstract":"This paper proposes a novel method that utilizes inter-view correlation in order to reduce the bitrate required to represent the prediction mode including the intra prediction direction, the motion vectors, the reference picture indexes, and the sub-block partitioning. The proposed method derives these kinds of macroblock (MB) information at the decoder side by comparing each image signal predictor with the view synthesis pictures. To accurately evaluate the predictors, we also propose a measure that minimizes the impact of errors in the depth information and view-dependent image signals. Experiments show that the proposed method reduces the bitrate by up to 25% relative to the international multiview video coding standard, and about 20% relative to the conventional view synthesis prediction method.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128662301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Experimental service of 3DTV cable broadcasting using dual HD streams 采用双高清流的3DTV有线广播实验服务
Dong-Hoon Lee, Hong-Ik Kim, Joonyoung Jung, Woongshik You
{"title":"Experimental service of 3DTV cable broadcasting using dual HD streams","authors":"Dong-Hoon Lee, Hong-Ik Kim, Joonyoung Jung, Woongshik You","doi":"10.1109/3DTV.2011.5877235","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877235","url":null,"abstract":"An experimental service of 3DTV cable broadcasting was made by delivering stereoscopic 3D video to the users in dual HD stream format. Also, an experimental attempt was made to provide backward compatibility for conventional 2D Set-top Boxes (STB) receiving and displaying only one stream of stereoscopic 3D video signal called reference video. For the experimental service, stereoscopic 3D video signals are encoded with 35 Mbps and transmitted over single broadcasting channel by using the existing cable headend and transmission systems. Prototypes of the software-based 3D encoder and 3D STB were also used for the experimental service. Through the experimental broadcasting service, it was possible to provide HD 3DTV broadcasting service under legacy cable environment and to support backward compatibility for the conventional 2D STBs to receive HD broadcasting by using the reference video. A survey revealed that about 60% of the viewers answered the image quality of stereoscopic 3D video in dual HD stream format was better than that in single HD stream format.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"78 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121669845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信