H. Sasaki, N. Okaichi, H. Watanabe, M. Kano, M. Kawakita, T. Mishina
{"title":"Color moiré reduction and resolution enhancement technique for integral three-dimensional display","authors":"H. Sasaki, N. Okaichi, H. Watanabe, M. Kano, M. Kawakita, T. Mishina","doi":"10.1109/3DTV.2017.8280402","DOIUrl":"https://doi.org/10.1109/3DTV.2017.8280402","url":null,"abstract":"In the case of using a flat-panel display for displaying elemental images in an integral three-dimensional (3D) display, the color moiré on the 3D image caused by the sub-pixel structure for colorization becomes a visual obstacle. Generally, a method of displaying an elemental image through a diffusion screen, or a defocusing method of the elemental image from the focal length position of the elemental lens have been used for color moiré reduction. However, on the other hand, this lowers the resolution of the elemental image, which is one of the causes of lowering the depth reproducibility of the 3D image. We developed a technology to reduce color moiré, while improving the resolution of reconstructed 3D images by multiple 3D display systems.","PeriodicalId":279013,"journal":{"name":"2017 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130076689","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Lossless compression of subaperture images using context modeling","authors":"I. Schiopu, M. Gabbouj, A. Gotchev, M. Hannuksela","doi":"10.1109/3DTV.2017.8280403","DOIUrl":"https://doi.org/10.1109/3DTV.2017.8280403","url":null,"abstract":"The paper proposes a method for lossless compression of subaperture image stacks obtained by rectifying light-field images captured by a plenoptic camera. We exploit the similarities between two subaperture images using a predictive coding algorithm, where the current view is predicted from one reference view. Context modeling is the main technique used to reduce the image file size. A suitable image segmentation and a template context are used by the context tree algorithm for encoding up to the smallest detail in each subaperture image. Entropy coding is configured by a residual analysis module. The results show improved performance compared to the state-of-the-art encoders.","PeriodicalId":279013,"journal":{"name":"2017 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"132 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120953414","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Eleftherios Anastasovitis, D. Ververidis, S. Nikolopoulos, Y. Kompatsiaris
{"title":"Digiart: Building new 3D cultural heritage worlds","authors":"Eleftherios Anastasovitis, D. Ververidis, S. Nikolopoulos, Y. Kompatsiaris","doi":"10.1109/3DTV.2017.8280406","DOIUrl":"https://doi.org/10.1109/3DTV.2017.8280406","url":null,"abstract":"The use of serious game (SG) for learning is now further extended in cultural heritage (CH). SGs simulate the real world through the application of 3D models that initiates interaction, imagination and trigger the immersion of the user. The design and development of three SGs that gamify three case studies of european CH are presented, in the context of Digiart European Union funded project. The first case study concerns the Scladina cave in Belgium. The second SG simulates an anthropology virtual museum for Liverpool John Moores University in United Kingdom. The third case study communicates with the public, the Palace of Aigai in Greece. Pioneering techniques have been used for generating accurate 3D models such as archaeological sites and artifacts. The embodiment of these assets into an interactive and fully immersive virtual environment of a game engine is the subject of this report.","PeriodicalId":279013,"journal":{"name":"2017 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132669093","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Mixed-resolution HEVC based multiview video codec","authors":"Bruhanth Mallik, A. S. Akbari, A. Kor","doi":"10.1109/3DTV.2017.8280426","DOIUrl":"https://doi.org/10.1109/3DTV.2017.8280426","url":null,"abstract":"Studies have shown that mixed resolution based video codecs, also known as asymmetric spatial inter/intra view video codecs are successful in efficiently coding videos for low bitrate transmission. In this paper a HEVC based spatial resolution scaling type of mixed resolution coding model for frame interleaved multiview videos is presented. The proposed codec is designed such that the information in intermediate frames of the center and neighboring views are down-sampled, while the frames still retaining the original size. The codec's reference frames structure is designed to efficiently encode frame interleaved multiview videos using a HEVC based mixed resolution codec. The multiview test video sequences were coded using the proposed codec and the standard MV-HEVC. Results show that the proposed codec gives significantly higher coding performance over the MV-HEVC codec at low bitrates.","PeriodicalId":279013,"journal":{"name":"2017 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115819133","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Xichen Zhou, B. Desai, Charalambos (Charis) Poullis
{"title":"Automatic 2D to stereoscopic video conversion for 3D TVS","authors":"Xichen Zhou, B. Desai, Charalambos (Charis) Poullis","doi":"10.1109/3DTV.2017.8280410","DOIUrl":"https://doi.org/10.1109/3DTV.2017.8280410","url":null,"abstract":"In this paper we present a novel technique for automatically converting 2D videos to stereoscopic. Uniquely, the proposed approach leverages the strengths of Deep Learning to address the complex problem of depth estimation from a single image. A Convolutional Neural Network is trained on input RGB images and their corresponding depths maps. We reformulate and simplify the process of generating the second camera's depth map and present how this can be used to render an anaglyph image. The anaglyph image was used for demonstration only because of the easy and wide availability of red/cyan glasses however, this does not limit the applicability of the proposed technique to other stereo forms. Finally, we present preliminary results and discuss the challenges.","PeriodicalId":279013,"journal":{"name":"2017 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125787235","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Karakostas, D. Ververidis, S. Nikolopoulos, Y. Kompatsiaris
{"title":"SpAtiAL: A sensor based framework to support affective learning","authors":"A. Karakostas, D. Ververidis, S. Nikolopoulos, Y. Kompatsiaris","doi":"10.1109/3DTV.2017.8280400","DOIUrl":"https://doi.org/10.1109/3DTV.2017.8280400","url":null,"abstract":"The objective of this paper is to present SpAtiAL, a new framework to design and develop an intelligent and personalized affective educational environment that will be able to support learners in multiple education settings. The sensors-based environment will support adaptively a) learners based on their enriched affective models and b) teachers in order to orchestrate the learning procedures more efficiently. In SpAtiAL we propose a set of Electromyogram (EMG) based features such as muscles total pressure, flexors pressure, tensors pressure, and gesture stiffness, for the purpose of identifying differences in students' affective state. In identifying these EMG-based features we have developed a tool for visualizing in real-time the signals generated from a Myo sensor along with the muscle activation level in 3D space.","PeriodicalId":279013,"journal":{"name":"2017 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115011925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Read-the-game skill evaluation by analyzing head orientation in immersive VR","authors":"César Daniel Rojas Ferrer, I. Kitahara, Y. Kameda","doi":"10.1109/3DTV.2017.8280415","DOIUrl":"https://doi.org/10.1109/3DTV.2017.8280415","url":null,"abstract":"This paper proposes a Virtual Reality (VR) system that allows both players and coaches to measure Active Visual Exploratory Activity patterns (AVEA). The main purpose of the system is to analyze the ability to “read the game” of soccer players under pressure. By making use of Head Mounted Display (HMD) technology and its head tracking capabilities, the users experience an in-game situation. They are set in the place of a defender with ball possession before making a passing decision while being pressured by a rival player. Therefore, being forced to make quick decisions based on visual information. By extracting the motion tracking data of the HMD, the gazing direction of the user is saved and divided in three zones of interest. A trial test was carried (N=10) to evaluate the system and the IPQ questionnaire was applied post session for analyzing AVEA performance and presence relationship.","PeriodicalId":279013,"journal":{"name":"2017 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128588481","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Manolis Vasileiadis, Dimitrios Giakoumis, S. Malassiotis, I. Kostavelis, D. Tzovaras
{"title":"Body-part tracking from partial-view depth data","authors":"Manolis Vasileiadis, Dimitrios Giakoumis, S. Malassiotis, I. Kostavelis, D. Tzovaras","doi":"10.1109/3DTV.2017.8280408","DOIUrl":"https://doi.org/10.1109/3DTV.2017.8280408","url":null,"abstract":"This paper presents a high-accuracy body-part tracking algorithm, capable of achieving efficient human motion analysis from partial view depth-data, suitable for deployment in real-life applications. The algorithm uses a consumer-grade depth camera for data input and combines a discriminative body part estimator along with a generative tracker, utilizing a realistic human body model, in order to track individual body limbs in short camera-distance, partial-view scenarios. Additionally, a shape adaptation feature is also introduced in order to further morph the human model based on the observations. The implementation is tested in a lower-body limbs tracking scenario, achieving promising accuracy and performance on consumer-grade hardware. Moreover, a lower-body motion dataset is also provided, consisting of 16 real-world sequences using automatic ground-truth annotations from a commercial motion capture system.","PeriodicalId":279013,"journal":{"name":"2017 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130640834","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Accurate multi-view stereo by selective expansion","authors":"Hu Tian, Fei Li","doi":"10.1109/3DTV.2017.8280405","DOIUrl":"https://doi.org/10.1109/3DTV.2017.8280405","url":null,"abstract":"We present a multi-view stereo method for robust and efficient dense modeling based on selective expansion. The core is that the proposed method progressively refines pixel depths by expanding pixels that are selected with small photo-consistency costs to their neighbors, which is more efficient than conventional PatchMatch methods. Besides, a depth refinement process including mapping and filtering against neighboring frames is used to further improve the accuracy of estimated depths. In contrast to previous methods, our method has low computational complexity and doesn't need any optimization algorithm. The accuracy of proposed method is evaluated quantitatively and qualitatively on benchmark data.","PeriodicalId":279013,"journal":{"name":"2017 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123149195","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Hidehiko Shishido, Aoi Harazaki, Y. Kameda, I. Kitahara
{"title":"Smoothly switching method of asynchronous multi-view videos using frame interpolation","authors":"Hidehiko Shishido, Aoi Harazaki, Y. Kameda, I. Kitahara","doi":"10.1109/3DTV.2017.8280409","DOIUrl":"https://doi.org/10.1109/3DTV.2017.8280409","url":null,"abstract":"This paper proposes a method that generates viewpoint smooth switching by reducing the flickering artifact observed at bullet-times generated from asynchronous multi-view videos using frame interpolation processing. When we asynchronously capture multi-view videos of an object moving at high velocity, deviations occur in the observed position at the bullet-times. We apply a frame interpolation technique to reduce the problem. By selecting suitable interpolated images that produce the smallest movement of the subject's observed position, we smoothly generate viewpoint switched bullet-time video.","PeriodicalId":279013,"journal":{"name":"2017 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133934251","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}