{"title":"Free-View TV watermark selection based on the distribution characteristics","authors":"Evlambios E. Apostolidis, G. Triantafyllidis","doi":"10.1109/3DTV.2009.5069639","DOIUrl":"https://doi.org/10.1109/3DTV.2009.5069639","url":null,"abstract":"In Free-View Television (FTV), the user can interactively control the viewpoint and generate new arbitrary views of a dynamic scene from any 3D position. The new views might be recorded and misused. Therefore the problem of copyright and copy protection in FTV should be solved. Among many alternative rights management methods, the copyright problem for visual data can be approached by means of embedding hidden imperceptible information, called watermark, into the image and video content. But this approach differs from the simple watermarking technique, since watermark in FTV should not only be resistant to common video processing and multi-view video processing operations, it should also be easily extracted from a generated video of an arbitrary view. In this paper, we focus on the evaluation of the performance of several watermarks according to their distribution characteristics in order to survive in the new generated arbitrary views of FTV.","PeriodicalId":230128,"journal":{"name":"2009 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125940064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Temporally consistent layer depth ordering via pixel voting for pseudo 3D representation","authors":"Engin Turetken, A. Alatan","doi":"10.1109/3DTV.2009.5069679","DOIUrl":"https://doi.org/10.1109/3DTV.2009.5069679","url":null,"abstract":"A new region-based depth ordering algorithm is proposed based on the segmented motion layers with affine motion models. Starting from an initial set of layers that are independently extracted for each frame of an input sequence, relative depth order of every layer is determined following a bottom-to-top approach from local pair-wise relations to a global ordering. Layer sets of consecutive time instants are warped in two opposite directions in time to capture pair-wise occlusion relations of neighboring layers in the form of pixel voting statistics. Global depth order of layers is estimated by mapping the pair-wise relations to a directed acyclic graph and solving the longest path problem via a breadth-first search strategy. Temporal continuity is enforced both at the region segmentation and depth ordering stages to achieve temporally coherent layer support maps and depth order relations. Experimental results show that the proposed algorithm yields quite promising results even on dynamic scenes with multiple motions.","PeriodicalId":230128,"journal":{"name":"2009 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126625343","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Field of view expansion for 3-D holographic display using a single spatial light modulator with scanning reconstruction light","authors":"Rick H-Y Chen, T. Wilkinson","doi":"10.1109/3DTV.2009.5069673","DOIUrl":"https://doi.org/10.1109/3DTV.2009.5069673","url":null,"abstract":"Increasing the field of view of a holographic display while maintaining adequate image size is a difficult task. To address this problem, we designed a system that tessellates several sub-holograms into one large hologram at the output. The sub-holograms we generate is similar to a kinoform but without the paraxial approximation during computation. The sub-holograms are loaded onto a single spatial light modulator consecutively and relayed to the appropriate position at the output through a combination of optics and scanning reconstruction light. We will review the method of computer generated hologram and describe the working principles of our system. Results from our proof-of-concept system are shown to have an improved field of view and reconstructed image size.","PeriodicalId":230128,"journal":{"name":"2009 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video","volume":"2016 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127357230","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
T. Colleu, L. Morin, C. Labit, S. Pateux, R. Balter
{"title":"Compact quad-based representation for 3D video","authors":"T. Colleu, L. Morin, C. Labit, S. Pateux, R. Balter","doi":"10.1109/3DTV.2009.5069648","DOIUrl":"https://doi.org/10.1109/3DTV.2009.5069648","url":null,"abstract":"The context of this study is 3D video. Starting from a sequence of multi-view video plus depth (MVD) data, the proposed quad-based representation takes into account, in a unified manner, different issues such as compactness, compression, and intermediate view synthesis. The representation is obtained into two steps. Firstly, a set of 3D quads is extracted by using a quadtree decomposition of the depth maps. Secondly, a selective elimination of the quads is performed in order to reduce inter-view redundancies and thus provide a compact representation. Experiments on two real sequences show good quality results at the rendering stage and a small data overload compared to mono-view video.","PeriodicalId":230128,"journal":{"name":"2009 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121746985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Quality comparison and acceleration for digital hologram generation method based on segmentation","authors":"Hoonjong Kang, F. Yaras, L. Onural","doi":"10.1109/3DTV.2009.5069661","DOIUrl":"https://doi.org/10.1109/3DTV.2009.5069661","url":null,"abstract":"A holographic fringe pattern generation methods is based on Fraunhofer diffraction and subsequent segmentation and approximation of the fringe pattern. Several modifications of the original algorithm are already proposed to improve the quality of reconstructions. We compare the quality of to the reconstructed images from different versions of this algorithm by taking the reconstructions from the Fresnel hologram as a reference. Since, there is not any generally accepted objective quality assessment method for such reconstructions, we used some experimental methods such as intensity spread over the reconstructed images, total noise power, and peak-signal-to-noise for comparison. Then we chose the best performing algorithm in terms of ireconstruction quality, and developed a GPU-based implementation to accelerate the computation speed. The quality of the resultant reconstructions is comparable to reconstructions from Fresnel holograms; much higher speed is achieved due to multi-GPU implemetation.","PeriodicalId":230128,"journal":{"name":"2009 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122155611","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
O. Moslah, A. Valles-Such, V. Guitteny, Serge Couvet, Sylvie Philipp-Foliguet
{"title":"Accelerated multi-view stereo using parallel processing capababilities of the GPUS","authors":"O. Moslah, A. Valles-Such, V. Guitteny, Serge Couvet, Sylvie Philipp-Foliguet","doi":"10.1109/3DTV.2009.5069678","DOIUrl":"https://doi.org/10.1109/3DTV.2009.5069678","url":null,"abstract":"This paper presents an accelerated implementation of a multi-view stereo pipeline using parallel processing capababilities of the GPUs. Our system takes as input a set of calibrated photographs and produces a textured 3D mesh of the scene. The pipeline is divided into three parts: dense stereo matching, multi-view correspondence linking and 3D model generation. First, we use a combined vertical aggregation and dynamic programming (DP) scheme to produce disparity maps between pairs of photographs. Then, the depth maps are computed using a multi-view correspondence linking algorithm. Finally, we use a Delaunay triangulation algorithm and texture mapping to produce the 3D model of the scene.","PeriodicalId":230128,"journal":{"name":"2009 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128357634","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Voxel based annealed particle filtering for markerless 3D articulated motion capture","authors":"C. Canton-Ferrer, J. Casas, M. Pardàs","doi":"10.1109/3DTV.2009.5069645","DOIUrl":"https://doi.org/10.1109/3DTV.2009.5069645","url":null,"abstract":"This paper presents a view-independent approach to markerless human motion capture in low resolution sequences from multiple calibrated and synchronized cameras. Redundancy among cameras is exploited to generate a 3D voxelized representation of the scene and a human body model (HBM) is introduced towards analyzing these data. An annealed particle filtering scheme where every particle encodes an instance of the pose of the HBM is employed. Likelihood between particles and input data is performed using occupancy and surface information and kinematic constrains are imposed in the propagation step towards avoiding impossible poses. Test over the HumanEva annotated dataset yield quantitative results showing the effectiveness of the proposed algorithm.","PeriodicalId":230128,"journal":{"name":"2009 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134229312","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Evaluation of stereo video coding schemes for mobile devices","authors":"A. Aksay, G. Akar","doi":"10.1109/3DTV.2009.5069664","DOIUrl":"https://doi.org/10.1109/3DTV.2009.5069664","url":null,"abstract":"Mobile devices such as mobile phones, personal digital assistants and personal video/game players are somehow converging and getting more powerful, thus enabling 3D mobile devices a reality. In order to store or transmit stereo video in these devices, coding techniques from both monoscopic video coding and multi-view video coding can be used. In this work, we analyze the possible stereoscopic encoding schemes for mobile devices. We have used rate-distortion curves for coding efficiency and decoding speed tests for decoder complexity. Depending on the processing power and memory of the mobile device, we concluded to use two of the settings used in our experiments.","PeriodicalId":230128,"journal":{"name":"2009 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114226745","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A novel 3D audio display system using radiated loudspeaker for future 3D multimodal communications","authors":"M. Katsumoto, Y. Yamakata, Toshiyuki Kimura","doi":"10.1109/3DTV.2009.5069683","DOIUrl":"https://doi.org/10.1109/3DTV.2009.5069683","url":null,"abstract":"It is very important to develop 3D audio display systems compatible with future 3D visual displays in order to realize future 3D multimodal communications systems such as 3D television. In this paper, two 3D audio display systems using the radiated loudspeaker, that were developed by the authors, are described. In one system, the spherical radiated loudspeaker, which consists of 26 loudspeaker units and displays the solo player, was used. In this study, the spherical radiated loudspeaker displays two types of solo players. In the other system, the rectangular radiated loudspeaker, which consists of 157 loudspeaker units and displays the ensemble, was used. In this study, the rectangular radiated loudspeaker displays the string quartet.","PeriodicalId":230128,"journal":{"name":"2009 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video","volume":"103 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114414461","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multimodal interaction of auditory spatial cues and passive observer movement in simulated self motion","authors":"W. Martens, S. Sakamoto, Yôiti Suzuki","doi":"10.1109/3DTV.2009.5069685","DOIUrl":"https://doi.org/10.1109/3DTV.2009.5069685","url":null,"abstract":"When the movement of an observer through a three-dimensional (3D) space is simulated via a multimodal display system, the synchrony of cues from different sensory modalities can help or hinder the creation of illusions of self motion. Stereoscopic stimuli provide particularly strong cues to self motion, through both binocular and motion disparity, but in the absence of these visual cues, movement cues can also be provided through spatial audio and through “suggestive” movement of an observer using a motion platform. Indeed, passive movement of observers watching 3DTV can suggest more dramatic movement in space, which can facilitate interpretation of auditory cues, reinforcing illusions of observer movement through a presented space. This multimodal interaction is important for successful presentation of 3D information in future 3DTV applications, since changes in listening position can result in apparent sound source motion relative to a fixed listening position, rather than supporting illusions of self motion through a static scene. In order to quantitatively measure such effects, a multimodal interaction study was initiated in which visual cues were eliminated so as to focus upon sensitivity to temporal synchrony between passive whole-body motion and auditory spatial information. For simple spatial trajectories of two sound sources passing by the observer's position, the relative timing of passive movement of the observer was manipulated to produce a range of intermodal delays. It was found that decreasing the peak velocity reached by the moving sound sources allowed observers to tolerate more easily the presented asynchronies between the timing of this peak and the peak in whole-body motion, especially when the peak in motion occurred earlier in time than the peak in sound source velocity.","PeriodicalId":230128,"journal":{"name":"2009 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115132059","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}