{"title":"Focus on visual rendering quality through content-based depth map coding","authors":"Emilie Bosc, M. Pressigout, L. Morin","doi":"10.1109/PCS.2010.5702448","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702448","url":null,"abstract":"Multi-view video plus depth (MVD) data is a set of multiple sequences capturing the same scene at different viewpoints, with their associated per-pixel depth value. Overcoming this large amount of data requires an effective coding framework. Yet, a simple but essential question refers to the means assessing the proposed coding methods. While the challenge in compression is the optimization of the rate-distortion ratio, a widely used objective metric to evaluate the distortion is the Peak-Signal-to-Noise-Ratio (PSNR), because of its simplicity and mathematically easiness to deal with such purposes. This paper points out the problem of reliability, concerning this metric, when estimating 3D video codec performances. We investigated the visual performances of two methods, namely H.264/MVC and Locally Adaptive Resolution (LAR) method, by encoding depth maps and reconstructing existing views from those degraded depth images. The experiments revealed that lower coding efficiency, in terms of PSNR, does not imply a lower rendering visual quality and that LAR method preserves the depth map properties correctly.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121273455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Shun-ichi Sekiguchi, Akira Minezawa, K. Sugimoto, A. Ichigaya, Kazuhisa Iguchi, Y. Shishikui
{"title":"A novel video coding scheme for Super Hi-Vision","authors":"Shun-ichi Sekiguchi, Akira Minezawa, K. Sugimoto, A. Ichigaya, Kazuhisa Iguchi, Y. Shishikui","doi":"10.1109/PCS.2010.5702498","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702498","url":null,"abstract":"We propose a novel video coding scheme targeting Super Hi-Vision (SHV) video sources. While it takes a conventional block-based MC + Transform hybrid coding approach that is suitable for hardware implementation of a SHV video codec, the proposed scheme achieved significant coding efficiency improvement by introducing several coding tools such as intra prediction and adaptive transform. According to our experimental analysis, the proposed scheme achieves significant bit-rate saving compared to the state-of-the-art AVC/H.264 high profile.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"91 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124422331","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Ishikawa, Hiroshi Sankoh, S. Naito, S. Sakazawa
{"title":"Efficient free viewpoint video-on-demand scheme realizing walk-through experience","authors":"A. Ishikawa, Hiroshi Sankoh, S. Naito, S. Sakazawa","doi":"10.1109/PCS.2010.5702452","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702452","url":null,"abstract":"This paper presents an efficient video-on-demand (VOD) scheme for free viewpoint television (FTV), and proposes a data format and its data generation method to provide a walkthrough experience. We employ a hybrid rendering approach to describe a 3D scene using 3D model data for objects and textures. However, conventional hybrid rendering methods such as multi-texturing include excessive redundancy in texture data and demand a great deal of bandwidth to transmit. In this paper we propose an efficient texture data format, which removes the redundancy due to occlusion of objects by employing an orthogonal projection image for each object. The additional advantage of the data format is great simplification at the server to choose the transmitted images that correspond to the requested viewpoint. Experiments using multiview real video sequences confirm that the proposed scheme can reduce the transmission of texture data by as much as 42% compared to the conventional scheme.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"101 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124797228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Md. Nazmul Haque, Moyuresh Biswas, M. Pickering, M. Frater
{"title":"An adaptive low-complexity global motion estimation algorithm","authors":"Md. Nazmul Haque, Moyuresh Biswas, M. Pickering, M. Frater","doi":"10.1109/PCS.2010.5702574","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702574","url":null,"abstract":"One important recent application of image registration has been in the estimation of global motion parameters for object-based video coding. A limitation of current global motion estimation approaches is the additional complexity of the gradient-descent optimization that is typically required to calculate the optimal set of global motion parameters. In this paper we propose a new low-complexity algorithm for global motion estimation. The complexity of the proposed algorithm is reduced by performing the majority of the operations in the gradient-descent optimization using logic operations rather than full-precision arithmetic operations. This use of logic operations means that the algorithm can be implemented much more easily in hardware platforms such as field programmable gate arrays (FPGAs). Experimental results show that the execution time for software implementations of the new algorithm is reduced by a factor of almost four when compared to existing fast implementations without any significant loss in registration accuracy.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"106 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115034438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Advanced inpainting-based macroblock prediction with regularized structure propagation in video compression","authors":"Yang Xu, H. Xiong","doi":"10.1109/PCS.2010.5702587","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702587","url":null,"abstract":"In this paper, we propose an optimized inpainting-based macroblock (MB) prediction mode (IP-mode) in the state-of-the-art H.264/AVC video compression engine, and investigate a natural extension of structured sparsity over the ordered Belief Propagation (BP) inference in inpainting-based prediction. The IP-mode is regularized by a global spatio-temporal consistency between the predicted content and the co-located known texture, and could be adopted in both Intra and Inter frames without redundant assistant information. It is solved by an optimization problem under Markov Random Field (MRF), and the structured sparsity of the predicted macroblock region is inferred by tensor voting projected from the decoded regions to tune the priority of message scheduling in BP with a more convergent manner. Rate-distortion optimization is maintained to select the optimal mode among the inpainting-based prediction (IP-), the intra-, and inter-modes. Compared to the existing prediction modes in H.264/AVC, the proposed inpainting-based prediction scheme is validated to achieve a better R-D performance for homogeneous visual patterns and behave a more robust error resilience capability with an intrinsic probabilistic inference.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115202528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
S. Wahl, Zhe Wang, Chensheng Qiu, M. Wróblewski, L. Rockstroh, S. Simon
{"title":"Memory-efficient parallelization of JPEG-LS with relaxed context update","authors":"S. Wahl, Zhe Wang, Chensheng Qiu, M. Wróblewski, L. Rockstroh, S. Simon","doi":"10.1109/PCS.2010.5702444","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702444","url":null,"abstract":"Many state-of-the-art lossless image compression standards feature adaptive error modelling. This, however, leads to data dependency loops of the compression scheme such that a parallel compression of neighboring pixels is not possible. In this paper, we propose a relaxation to the context update of JPEG-LS by delaying the update procedure, in order to achieve a guaranteed degree of parallelism with a negligible effect on the compression ratio. The lossless mode of JPEG-LS including the run-mode is considered. A descewing scheme is provided generating a bit-stream that preserves the order needed for the decoder to mimic the prediction in a consistent way. This system is memory efficient in a sense that no additional memory for the large context-set is needed.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122088327","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Enhanced video compression with region-based texture models","authors":"Fan Zhang, D. Bull","doi":"10.1109/PCS.2010.5702560","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702560","url":null,"abstract":"This paper presents a region-based video compression algorithm based on texture warping and synthesis. Instead of encoding whole images or prediction residuals after translational motion estimation, this algorithm employs a perspective motion model to warp static textures and uses a texture synthesis approach to synthesise dynamic textures. Spatial and temporal artefacts are prevented by an in-loop video quality assessment module. The proposed method has been integrated into an H.264 video coding framework. The results show significant bitrate savings, up to 55%, compared with H.264, for similar visual quality.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122223738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Eugen Wige, Gilbert Yammine, P. Amon, A. Hutter, André Kaup
{"title":"Analysis of in-loop denoising in lossy transform coding","authors":"Eugen Wige, Gilbert Yammine, P. Amon, A. Hutter, André Kaup","doi":"10.1109/PCS.2010.5702584","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702584","url":null,"abstract":"When compressing noisy image sequences, the compression efficiency is limited by the noise amount within these image sequences as the noise part cannot be predicted. In this paper, we investigate the influence of noise within the reference frame on lossy video coding of noisy image sequences. We estimate how much noise is left within a lossy coded reference frame. Therefore we analyze the transform and quantization step inside a hybrid video coder, specifically H.264/AVC. The noise power after transform, quantization, and inverse transform is calculated analytically. We use knowledge of the noise power within the reference frame in order to improve the inter frame prediction. For noise filtering of the reference frame, we implemented a simple denoising algorithm inside the H.264/AVC reference software JM15.1. We show that the bitrate can be decreased by up to 8.1% compared to the H.264/AVC standard for high resolution noisy image sequences.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117314785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Free viewpoint image generation with super resolution","authors":"Norishige Fukushima, Y. Ishibashi","doi":"10.1109/PCS.2010.5702462","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702462","url":null,"abstract":"In this paper, we propose a method of free viewpoint image generation with super resolution. In the conventional approaches, such as nearest neighbor and linear interpolation, the synthetic image on zoomed virtual view tends to have low resolution, because the reference images do not have enough textures. To overcome this problem, we reconstruct the image with super resolution. Super resolution can generate higher image resolution than the input image one, and then we combine super resolution with free viewpoint image generation. In the experiment, we use a camera array which contains 11 × 11 aligned cameras and use 4 × 4 cameras subset per pixel to reconstruct image by means of super resolution. The experimental results show that synthesized image in the effective range has about 4.5 dB higher PSNR than ones created by the nearest neighbor and 2.5 dB higher than ones created by the linear interpolation.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122276141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A new hybrid parallel intra coding method based on interpolative prediction","authors":"Cui Wang, Akira Kubota, Y. Hatori","doi":"10.1109/PCS.2010.5702521","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702521","url":null,"abstract":"The hybrid coding method combining the predictive coding with the orthogonal transformation and the quantization is mainly used recently. This paper proposes a new hybrid parallel Intra Coding based on interpolative prediction which uses correlations between neighboring pixels, including non-causal pixels. In order to get high prediction performance, the optimal quantizing scheme, which is used to cancel the error that expands when decoding, is used. Furthermore, a new type of block shape, which enables parallel coding, is proposed to simplify the processing of interpolative prediction. The result of comparison between proposed method and intra coding method in H.264 shows that the PSNR of proposed technique achieves 1 dB to 4 dB improvement in Luminance, especially for image with more details.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127985664","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}