{"title":"An Intra-Affine Current Picture Referencing Mode for Screen Content Coding in VVC","authors":"Jian Cao, Zhengren Li, Fan Liang, Jun Wang","doi":"10.1109/PCS48520.2019.8954509","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954509","url":null,"abstract":"With the rapid development of emerging applications, screen content coding (SCC) is playing a more and more important role. Intra block copy (IBC), as a new tool for SCC, is proved to be efficient when there are many repeating or similar areas within the same picture. However, IBC is based on translational motion model, which may not work well for some blocks with complicated movement content, such as rotation and zoom. In this paper, a new intra-affine current picture referencing mode is proposed. In this new mode, non-translational motion model (affine model) is introduced to intra prediction for SCC to improve coding efficiency. First, candidate set of initial block-affine vectors (BVAffis) is established by a newly designed method. Then, those initial BVAffis are updated through iterative searching algorithm. Moreover, compatibility checking is applied. Compared to VTM3.0, the proposed new mode can achieve 2.53%, 2.47%, and 2.48% BD-rate saving on average for Y, U and V respectively for SCC test sequences. The draft about the new mode (JVET-O0682) was submitted in the 15th JVET meeting.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114630788","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Wieckowski, Jackie Ma, Valeri George, H. Schwarz, D. Marpe, T. Wiegand
{"title":"Generalized binary splits: A versatile partitioning scheme for block-based hybrid video coding","authors":"A. Wieckowski, Jackie Ma, Valeri George, H. Schwarz, D. Marpe, T. Wiegand","doi":"10.1109/PCS48520.2019.8954564","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954564","url":null,"abstract":"Block partitioning is the basis of every modern hybrid video coding standard. It specifies how the video pictures can be subdivided into blocks for prediction and residual coding. In H.265/HEVC, quad-tree partitioning is one of the key technologies allowing for flexible mode allocation and providing a substantial part of the gains over H.264/AVC. The current draft of the upcoming standard Versatile Video Coding (VVC) provides over 30% bit-rate savings over HEVC and almost one third of the gain is achieved by using a more flexible partitioning scheme than the quad-tree partitioning used in HEVC. In this paper, we describe a partitioning concept that generalizes many of the ideas developed during the exploration and early standardization phase of VVC. In fact, our method includes the VVC partitioning as well as many other state-of-the-art methods. The proposed method can be implemented in a fully configurable design. For instance, it can be configured to match the performance of VTM-1.0 at much faster runtime (69%) or it can be configured to obtain additional bit-rate savings of up to 3% by exploiting additional degrees of freedom.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114201037","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"B-DRRN: A Block Information Constrained Deep Recursive Residual Network for Video Compression Artifacts Reduction","authors":"Hoang Man Trinh, Jinjia Zhou","doi":"10.1109/PCS48520.2019.8954521","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954521","url":null,"abstract":"Although the video compression ratio nowadays becomes higher, the video coders such as H.264/AVC, H.265/HEVC, H.266/VVC always suffer from the video artifacts. In this paper, we design a neural network to enhance the quality of the compressed frame by leveraging the block information, called B-DRRN (Deep Recursive Residual Network with Block information). Firstly, an extra network branch is designed for leveraging the block information of the coding unit (CU). Moreover, to avoid a great increase in the network size, Recursive Residual structure and sharing weight techniques are applied. We also conduct a new large-scale dataset with 209,152 training samples. Experimental results show that the proposed B-DRRN can reduce 6.16% BD-rate compared to HEVC standard. After efficiently adding an extra network branch, this work can improve the performance of the main network without increasing any memory for storing.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114471405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Christian Rudat, Christian R. Helmrich, J. Lainema, Tung Nguyen, H. Schwarz, D. Marpe, T. Wiegand
{"title":"Inter-Component Transform for Color Video Coding","authors":"Christian Rudat, Christian R. Helmrich, J. Lainema, Tung Nguyen, H. Schwarz, D. Marpe, T. Wiegand","doi":"10.1109/PCS48520.2019.8954496","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954496","url":null,"abstract":"In natural digital images and videos, correlations between color components can be observed. These correlations can be exploited to achieve additional coding gain in modern block-based hybrid video coding. To this end, we propose the use of a block-wise, rotational inter-component transform (ICT) applied to the two residual chroma signals that result from conventional intra or inter-picture prediction. Different ICT parameterizations in terms of number and quantization of the rotational angles as well as resulting components signaled in the coded bitstream are investigated. An implementation into the currently developed Versatile Video Coding (VVC) reference software provides average bitrate savings of up to 0.7% (All Intra configuration) with negligible increases in implementation complexity and runtime. Our proposal has been adopted into the VVC draft specification text.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124945037","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jennifer Rasch, Jonathan Pfaff, H. Schwarz, D. Marpe, T. Wiegand
{"title":"Comparison between the Diffusion and the ADMM Filter and Combined Results","authors":"Jennifer Rasch, Jonathan Pfaff, H. Schwarz, D. Marpe, T. Wiegand","doi":"10.1109/PCS48520.2019.8954493","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954493","url":null,"abstract":"In this paper we compare two different methods to filter prediction signals in video coding: The anisotropic diffusion filter modeled using a (modified) heat equation and the ADMM (Alternating Direction Method of Multipliers) filter using directional total variation. The two methods are based on the same idea, namely using the structures in the underlying prediction signal to control the direction of the smoothing. It is shown here, that mathematically, in a continuous setting for a certain choice of parameters, the solutions of the two problem formulations are the same.Both filters are compared in the setting of a software based on HEVC (High Efficiency Video Coding). While the results for the diffusion filter method involve the testing of four different filters of uniform and signal adaptive type, the ADMM filter results feature only one type of filter configuration that can be en- or disabled in a block-wise manner. Despite the fact that both methods rely on the same idea, it is illustrated that the overall gains of the combination are higher than the RD gains of the individual tools. In particular, it is demonstrated that this is still the case if the filters are combined in an exclusive way. This proves that there is an additional value for each of the filter methods.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132228019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Baptiste Hériard-Dubreuil, Irene Viola, T. Ebrahimi
{"title":"Light field compression using translation-assisted view estimation","authors":"Baptiste Hériard-Dubreuil, Irene Viola, T. Ebrahimi","doi":"10.1109/PCS48520.2019.8954495","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954495","url":null,"abstract":"Light field technology has recently been gaining traction in the research community. Several acquisition technologies have been demonstrated to properly capture light field information, and portable devices have been commercialized to the general public. However, new and efficient compression algorithms are needed to sensibly reduce the amount of data that needs to be stored and transmitted, while maintaining an adequate level of perceptual quality. In this paper, we propose a novel light field compression scheme that uses view estimation to recover the entire light field from a small subset of encoded views. Experimental results on a widely used light field dataset show that our method achieves good coding efficiency with average rate savings of 54.83% with respect to HEVC.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"117 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124187396","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Mischa Siekmann, S. Bosse, H. Schwarz, D. Marpe, T. Wiegand
{"title":"Data-driven Optimization of Row-Column Transforms for Block-Based Hybrid Video Compression","authors":"Mischa Siekmann, S. Bosse, H. Schwarz, D. Marpe, T. Wiegand","doi":"10.1109/PCS48520.2019.8954516","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954516","url":null,"abstract":"In state-of-the-art video compression residual coding is done by transforming the prediction error signals into a less correlated representation and performing the quantization and entropy coding in the transform domain. For complexity reasons usually separable transforms are used. A more flexible transform structure is given by row-column transforms, which apply a separate transform to each row and each column of a signal block. This paper describes a method for training such structured transforms by maximizing the data likelihood under a parameterized probabilistic model with a compelled structure. An explicit model is derived for the case of row-column transforms and its efficiency is demonstrated in the application of video compression. It is shown that trained row-column transforms achieve almost the same coding gain as unconstrained KLTs when applied as secondary transforms, while the encoder and decoder runtime are the same as in the separable transform case.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124188146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Shuyuan Zhu, Zhiying He, Xiandong Meng, Guanghui Liu, B. Zeng
{"title":"Multiple Description Image Coding Based on Compression-Guided Optimization","authors":"Shuyuan Zhu, Zhiying He, Xiandong Meng, Guanghui Liu, B. Zeng","doi":"10.1109/PCS48520.2019.8954534","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954534","url":null,"abstract":"In this paper, we design a new multiple description coding scheme for image signals based on our proposed compression-guided optimization. Firstly, we propose a compression-constrained adaptive filtering method to produce two descriptions for the source image, where the proposed filtering algorithm works not only to guarantee a high-quality side decoding but also make a high-efficient central decoding. Secondly, we design a compression-dependent deblocking algorithm based on the transform coefficients which are decoded from both descriptions to improve the performance for the cental decoding. Experimental results demonstrate that our proposed method achieves impressive performance gains when it is applied to image signals.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126892368","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Quality Assessment of Stereoscopic 360-degree Images from Multi-viewports","authors":"Jiahua Xu, Ziyuan Luo, Wei Zhou, Wenyuan Zhang, Zhibo Chen","doi":"10.1109/PCS48520.2019.8954555","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954555","url":null,"abstract":"Objective quality assessment of stereoscopic panoramic images becomes a challenging problem owing to the rapid growth of 360-degree contents. Different from traditional 2D image quality assessment (IQA), more complex aspects are involved in 3D omnidirectional IQA, especially unlimited field of view (FoV) and extra depth perception, which brings difficulty to evaluate the quality of experience (QoE) of 3D omnidirectional images. In this paper, we propose a multi-viewport based full-reference stereo 360 IQA model. Due to the freely changeable viewports when browsing in the head-mounted display, our proposed approach processes the image inside FoV rather than the projected one such as equirectangular projection (ERP). In addition, since overall QoE depends on both image quality and depth perception, we utilize the features estimated by the difference map between left and right views which can reflect disparity. The depth perception features along with binocular image qualities are employed to further predict the overall QoE of 3D 360 images. The experimental results on our public Stereoscopic OmnidirectionaL Image quality assessment Database (SOLID) show that the proposed method achieves a significant improvement over some well-known IQA metrics and can accurately reflect the overall QoE of perceived images.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"209 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122823199","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}