Lin Du, Peng Qin, Jianping Song, Wenjuan Song, Yan Xu, Wei Zhou
{"title":"Immersive 3D user interface for 3D TVS","authors":"Lin Du, Peng Qin, Jianping Song, Wenjuan Song, Yan Xu, Wei Zhou","doi":"10.1109/3DTV.2011.5877182","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877182","url":null,"abstract":"This paper investigates some important visual factors of immersive 3D user interfaces on the latest 3D displays and studies how they can be applied for 3D TVs at home. The vergence-accommodation conflict in immersive 3D user interfaces is addressed in order to theoretically estimate some perceived parameters of the user interfaces under this constraint. To demonstrate the proposed immersive 3D user interfaces, a 3D TV browse and control demo is built and some stereo screenshot results are presented.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116555971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A fast image segmentation algorithm using color and depth map","authors":"Emanuele Mirante, M. Georgiev, A. Gotchev","doi":"10.1109/3DTV.2011.5877227","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877227","url":null,"abstract":"In this paper, a real-time image segmentation algorithm is presented. It utilizes both color and depth information retrieved from a multi-sensor capture system, which combines stereo camera pairs with time-of-flight range sensor. The algorithm targets low complexity and fast implementation which can be achieved through parallelization. Applications, such as immersive videoconferencing and lecturer segmentation for augmented reality lecture presentation, can benefit from the designed algorithm.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"122 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115961749","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Depth map generation for 2D-to-3D conversion by limited user inputs and depth propagation","authors":"Xi Yan, You Yang, Guihua Er, Qionghai Dai","doi":"10.1109/3DTV.2011.5877167","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877167","url":null,"abstract":"The quality of depth maps is crucial for 2D to 3D conversion, but high quality depth map generation methods are usually very time consuming. We propose an efficient semi-automatic depth map generation scheme based on limited user inputs and depth propagation. First, the original image is over-segmented. Then, the depth values of selected pixels and the approximate locations of T-junctions are specified by user inputs. The final depth map is obtained by depth propagation combining user inputs, color and edge information. The experimental results demonstrate that our scheme is satisfactory in terms of both accuracy and efficiency, and thus can be applied for high quality 2D to 3D video conversion.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125953540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Robust focal length estimation based on minimal solution method","authors":"Deqing Chen, Hang Shao, Qionghai Dai","doi":"10.1109/3DTV.2011.5877179","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877179","url":null,"abstract":"We present a new approach to estimate the focal length for camera calibration in multiview reconstruction. As a popular camera calibration approach, minimal solution method gives rise to a great number of focal-length estimates, from which generating an accurate one is of great significance. Our method concentrates on how to obtain an accurate estimate and is carried out in two steps: firstly, a norm constraint for the fundamental matrix is employed to prune the low-confidence focal-length candidates. Then the focal-length estimate is obtained with a robust focal-length estimation scheme, which consists of occurrence to probability transform, focal-length candidates resample and final estimation with expectation. Experimental results demonstrate that our method could obtain better estimate with both higher accuracy and higher stability than the state-of-the-art method.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134141730","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Self-adaptive normal estimation and position adjustment for MVS reconstruction","authors":"Yanjun Qian, Qionghai Dai, Guihua Er","doi":"10.1109/3DTV.2011.5877161","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877161","url":null,"abstract":"Generating a polygonal mesh model from the point cloud is a critical step of many state-of-art MVS reconstruction algorithms, and influences the accuracy and visual quality of the final results significantly. The normal estimation and position adjustment of each point is required for this procedure. We present a mathematical analysis of the normal estimation approach, and propose two hypotheses to determinate the accuracy and smoothness of the points in a local region. A multi-scale strategy is implemented to obtain a proper scale for each point. Then the according normal is calculated by PCA on this scale, and the positions can be optimized by combining the accurate neighboring normals. A 2D toy example proves that the proposed approach can adjust the noisy point to the right surface while preserving details. At last we show that our method can improve the quality of mesh models for real MVS reconstruction tasks.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131877861","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
P. Aflaki, M. Hannuksela, Jussi H. Hakala, J. Hakkinen, M. Gabbouj
{"title":"Estimation of subjective quality for mixed-resolution stereoscopic video","authors":"P. Aflaki, M. Hannuksela, Jussi H. Hakala, J. Hakkinen, M. Gabbouj","doi":"10.1109/3DTV.2011.5877171","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877171","url":null,"abstract":"In mixed-resolution (MR) stereoscopic video, one view is presented with a lower resolution compared with the other one; therefore, a lower bitrate, a reduced computational complexity, and a decrease in memory access bandwidth can be expected in coding. The human visual system is known to fuse left and right views in such a way that the perceptual visual quality is closer to that of the higher-resolution view. In this paper, a subjective assessment of mixed resolution (MR) stereoscopic videos is presented and the results are analyzed and compared with previous subjective tests presented in the literature. Three downsampling ratios 1/2, 3/8, and 1/4 were used to create lower-resolution views. Hence, the lower-resolution view had different spatial resolutions in terms of pixels per degree (PPD) for each downsampling ratio. It was discovered that the subjective viewing experience tended to follow a logarithmic function of the spatial resolution of the lower-resolution view measured in PPD. A similar behavior was also found from the results of an earlier experiment. Thus, the results suggest that the presented logarithmic function characterizes the expected viewing experience of MR stereoscopic video.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132829060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jae-Hyun Jung, Jisoo Hong, G. Park, Keehoon Hong, S. Min, Byoungho Lee
{"title":"Evaluation of perceived depth resolution in multi-view threedimensional display using depth image-based rendering","authors":"Jae-Hyun Jung, Jisoo Hong, G. Park, Keehoon Hong, S. Min, Byoungho Lee","doi":"10.1109/3DTV.2011.5877156","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877156","url":null,"abstract":"This paper describes the threshold of perception of depth resolution in multi-view three-dimensional display using depth image-based rendering (DIBR). In the pickup process, we capture the stereo image and center depth map in various depth resolutions using OpenGL. The multi-view images are generated by the depth image-based rendering of the stereo image and center depth map in different depth resolution. The synthesized view-images and simulation results are compared with regard to peak signal-to-noise ratio (PSNR), and the perception threshold of depth resolution is defined by the result of questionnaire with stereoscopic display.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130608328","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. O. Wildeboer, T. Yendo, M. Panahpour Tehrani, T. Fujii, M. Tanimoto
{"title":"Depth up-sampling for depth coding using view information","authors":"M. O. Wildeboer, T. Yendo, M. Panahpour Tehrani, T. Fujii, M. Tanimoto","doi":"10.1109/3DTV.2011.5877198","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877198","url":null,"abstract":"In 3DTV and free viewpoint imaging systems based on a view-plus-depth representation, depth compression is important for high-quality view synthesis. Several publications have proposed depth down-/up-sampling as part of the depth coding strategy. Recently, we proposed a depth up-sampling method which uses the high resolution view in the process of depth up-sampling. Actually, in 2007 Kopf et.al. already proposed a Joint Bilateral Upsampler(JBU), which uses a high resolution input image as prior in a joint bilateral filter for upscaling. In this paper we compare our previous method with the JBU approach, in the context of depth coding. Furthermore, we analyze the influence of view coding on the depth up-sampling result.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115205520","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
S. Shimizu, H. Kimata, Shiori Sugimoto, N. Matsuura
{"title":"Decoder-side macroblock information derivation for efficient multiview video plus depth map coding","authors":"S. Shimizu, H. Kimata, Shiori Sugimoto, N. Matsuura","doi":"10.1109/3DTV.2011.5877205","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877205","url":null,"abstract":"This paper proposes a novel method that utilizes inter-view correlation in order to reduce the bitrate required to represent the prediction mode including the intra prediction direction, the motion vectors, the reference picture indexes, and the sub-block partitioning. The proposed method derives these kinds of macroblock (MB) information at the decoder side by comparing each image signal predictor with the view synthesis pictures. To accurately evaluate the predictors, we also propose a measure that minimizes the impact of errors in the depth information and view-dependent image signals. Experiments show that the proposed method reduces the bitrate by up to 25% relative to the international multiview video coding standard, and about 20% relative to the conventional view synthesis prediction method.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128662301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Dong-Hoon Lee, Hong-Ik Kim, Joonyoung Jung, Woongshik You
{"title":"Experimental service of 3DTV cable broadcasting using dual HD streams","authors":"Dong-Hoon Lee, Hong-Ik Kim, Joonyoung Jung, Woongshik You","doi":"10.1109/3DTV.2011.5877235","DOIUrl":"https://doi.org/10.1109/3DTV.2011.5877235","url":null,"abstract":"An experimental service of 3DTV cable broadcasting was made by delivering stereoscopic 3D video to the users in dual HD stream format. Also, an experimental attempt was made to provide backward compatibility for conventional 2D Set-top Boxes (STB) receiving and displaying only one stream of stereoscopic 3D video signal called reference video. For the experimental service, stereoscopic 3D video signals are encoded with 35 Mbps and transmitted over single broadcasting channel by using the existing cable headend and transmission systems. Prototypes of the software-based 3D encoder and 3D STB were also used for the experimental service. Through the experimental broadcasting service, it was possible to provide HD 3DTV broadcasting service under legacy cable environment and to support backward compatibility for the conventional 2D STBs to receive HD broadcasting by using the reference video. A survey revealed that about 60% of the viewers answered the image quality of stereoscopic 3D video in dual HD stream format was better than that in single HD stream format.","PeriodicalId":158764,"journal":{"name":"2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"78 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121669845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}