2019 Picture Coding Symposium (PCS)最新文献

筛选
英文 中文
Unsymmetrical Quad-tree Partitioning for Audio Video coding Standard-3 (AVS-3) 音视频编码标准3 (AVS-3)的非对称四叉树分割
2019 Picture Coding Symposium (PCS) Pub Date : 2019-11-01 DOI: 10.1109/PCS48520.2019.8954558
Tianliang Fu, Kai Zhang, Li Zhang San, Hongbin Liu, Shanshe Wang, Siwei Ma
{"title":"Unsymmetrical Quad-tree Partitioning for Audio Video coding Standard-3 (AVS-3)","authors":"Tianliang Fu, Kai Zhang, Li Zhang San, Hongbin Liu, Shanshe Wang, Siwei Ma","doi":"10.1109/PCS48520.2019.8954558","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954558","url":null,"abstract":"Audio and Video coding Standard-3 (AVS-3) is an emerging next-generation video coding standard beyond AVS-2. The Quad-Tree/Binary-Tree (QTBT) plus Extend Quad-Tree (EQT) partitioning structure has been adopted in AVS3. In this paper, Unsymmetrical Quad-Tree (UQT) partitioning is proposed to explore the potential coding efficiency of partitioning further. Particularly, UQT partitioning splits a parent block into four sub-blocks asymmetrically with 1:4:2:1 or 1:2:4:1 ratio. Not only UQT partitioning can complement the current QTBT plus EQT partition structure but also can characterize the pattern of multifarious video content more effectively. Experiments are conducted on the HPM3.2 platform and the simulation results demonstrate that the proposed UQT partitioning can achieve 0.66%, 0.77% and 0.75% BD-bitrate savings for Y, Cb and Cr components on average under the random-access configuration, respectively.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128629553","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An Extended Skip Strategy for Inter Prediction 一种用于内部预测的扩展跳过策略
2019 Picture Coding Symposium (PCS) Pub Date : 2019-11-01 DOI: 10.1109/PCS48520.2019.8954532
Hao Tao, Li Yu, Zhuo Kuang, Hongkui Wang, Xiaofeng Huang
{"title":"An Extended Skip Strategy for Inter Prediction","authors":"Hao Tao, Li Yu, Zhuo Kuang, Hongkui Wang, Xiaofeng Huang","doi":"10.1109/PCS48520.2019.8954532","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954532","url":null,"abstract":"The High Efficiency Video Coding (HEVC) standard adopts inter prediction to eliminate temporal correlation between the successive frames. However, a large amount of bits need to be explicitly signaled in the bitstream to specify the motion information. In this paper, we propose an extended skip strategy to alleviate bit consumption for motion data during the inter prediction process. Specifically, before the current frame is encoded, an additional picture generated by a deep convolutional neural network (CNN) is introduced to inter prediction. Since the additional reference picture is more similar with the current frame, most blocks of this frame can be skipped in the coding process. Consequently, to further improve the compression, an extended skip strategy is designed, i.e., the current frame can be skipped in multi-levels, including frame-level and coding tree unit level (CTU-level). Moreover, the skip-level of the current frame is decided in the sense of rate-distortion optimization (RDO). The proposed algorithm is implemented on the HM-16.6 software and an average of 4.4% BD-rate gain has been achieved in the experiments, which indicates the superiority of the proposed method.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125237884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Perceptual Tolerance to Motion-To-Photon Latency with Head Movement in Virtual Reality 虚拟现实中头部运动对运动-光子延迟的感知容忍
2019 Picture Coding Symposium (PCS) Pub Date : 2019-11-01 DOI: 10.1109/PCS48520.2019.8954518
Minxia Yang, Jiaqi Zhang, Lu Yu
{"title":"Perceptual Tolerance to Motion-To-Photon Latency with Head Movement in Virtual Reality","authors":"Minxia Yang, Jiaqi Zhang, Lu Yu","doi":"10.1109/PCS48520.2019.8954518","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954518","url":null,"abstract":"Since Motion-To-Photon (MTP) latency is inevitable and can be perceived in virtual reality, quantifying perception of MTP latency becomes necessary. In this paper, we investigate perceptual tolerance to MTP latency, including perception threshold of the latency and user acceptance of delays above the threshold. It is affected by different head motion events, such as Motion-Static-Alternate (MSA, i.e., an acceleration or deceleration in one direction) and Motion-To-Reverse (MTR, i.e., a movement reverses direction). In each motion event, rotation angle and angular velocity also influence perception of MTP latency. Experimental results show that subjects are more intolerant of MTP latency in MTR than MSA. The latency perception threshold is about 23 ms when subjects turn their heads at the maximum speed of human limits. When the angular velocity decreases, the perception threshold increases. The maximum threshold is ~41 ms at 20 °/s in this study. Inversely proportional models are established to describe the relationship between threshold and angular velocity. Besides, MTP latency over the threshold is harder to be accepted with the rotation angle decreasing or the angular velocity increasing.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"50 10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127598802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Compression Performance of the Versatile Video Coding: HD and UHD Visual Quality Monitoring 通用视频编码的压缩性能:高清和超高清视觉质量监控
2019 Picture Coding Symposium (PCS) Pub Date : 2019-11-01 DOI: 10.1109/PCS48520.2019.8954562
N. Sidaty, W. Hamidouche, O. Déforges, P. Philippe, Jérôme Fournier
{"title":"Compression Performance of the Versatile Video Coding: HD and UHD Visual Quality Monitoring","authors":"N. Sidaty, W. Hamidouche, O. Déforges, P. Philippe, Jérôme Fournier","doi":"10.1109/PCS48520.2019.8954562","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954562","url":null,"abstract":"Video compression and content quality have become one of the most research topic in the recent years. Predominantly, trends obviously signpost that the video usage over the Internet is on the upsurge. Simultaneously, users’ requirement for enlarged resolution and higher quality is rising. Consequently, a huge effort has been made for video coding technologies and quality monitoring. In this paper, we present a subjective-based comparison as well as an objective measurement between the newest Versatile Video Coding (VVC) and the well-known High Efficiency Video Coding (HEVC) standards. Several videos of various content are selected as tested sequences. Both High Definition (HD) and Ultra High Definition (UHD) resolutions are used in this experiment. An extensive range of bit-rates from low to high bit-rates were selected. These sequences are encoded using both HEVC reference software (HM-16.2) and the latest reference software of VVC (VTM-5.0). Obtained results have shown that VVC outperforms consistently HEVC, for realistic bit rates and quality levels, in the range of 40% on the subjective scale. For the objective measurements, using PSNR, SSIM and VMAF as quality metrics, the quality enhancement of VVC over HEVC is ranging from 31% to 40%, depending on video content and spatial resolution.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131977030","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
A Data-Trained, Affine-Linear Intra-Picture Prediction in the Frequency Domain 频域数据训练的仿射线性图像内预测
2019 Picture Coding Symposium (PCS) Pub Date : 2019-11-01 DOI: 10.1109/PCS48520.2019.8954559
Michael Schäfer, Björn Stallenberger, Jonathan Pfaff, Philipp Helle, H. Schwarz, D. Marpe, T. Wiegand
{"title":"A Data-Trained, Affine-Linear Intra-Picture Prediction in the Frequency Domain","authors":"Michael Schäfer, Björn Stallenberger, Jonathan Pfaff, Philipp Helle, H. Schwarz, D. Marpe, T. Wiegand","doi":"10.1109/PCS48520.2019.8954559","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954559","url":null,"abstract":"This paper presents a data-driven training of affine- linear predictors which perform intra-picture prediction for video coding. The trained predictors use a single line of reconstructed boundary samples as input like the conventional intra prediction modes. For large blocks, the presented predictors initially transform the input samples via Discrete Cosine Transform. This allows to omit high frequency coefficients and consequently reduce the input dimension. The output is the result of a single matrix-vector multiplication and offset addition. Here, the predictors only construct certain coefficients in the frequency domain. The final prediction signal is then obtained by inverse transform. The coefficients of the prediction modes need to be stored in advance, requiring 0.273 MB of memory. The training employs a recursive block partitioning, where the loss function targets to approximate the bit-rate of the DCT-transformed block residuals. The obtained predictors are incorporated into the Versatile Video Coding Test Model 4. The authors report All- Intra bit-rate savings ranging from 0.7% to 2.0% across different resolutions in terms of the Bjøntegaard-Delta bit rate (BD-rate).","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124861480","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Impact of Video Streaming Delay on User Experience with Head-Mounted Displays 视频流延迟对头戴式显示器用户体验的影响
2019 Picture Coding Symposium (PCS) Pub Date : 2019-11-01 DOI: 10.1109/PCS48520.2019.8954527
Adam Grzelka, A. Dziembowski, Dawid Mieloch, O. Stankiewicz, J. Stankowski, M. Domański
{"title":"Impact of Video Streaming Delay on User Experience with Head-Mounted Displays","authors":"Adam Grzelka, A. Dziembowski, Dawid Mieloch, O. Stankiewicz, J. Stankowski, M. Domański","doi":"10.1109/PCS48520.2019.8954527","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954527","url":null,"abstract":"Delays in the delivery of immersive video to Head Mounted Devices (HMDs) are considered in the paper. The goal of this paper is to analyze the impact of video streaming latency on user experience with HMDs. The paper reports the results of the subjective quality assessment as a function of delay. The possible practical solutions are identified and, on this base, an experimental model of the considered systems is proposed. In order to properly test the influence of the streaming delay on the quality, the extensive subjective tests have been performed. Interesting conclusions have been drawn, showing that human acceptance of delay of translation of virtual point of viewing is much stronger than for delay of rotation of virtual viewing direction in immersive video. Finally, the influence of such observations on the system architectures is concluded.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116843796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Scalable Lossless Coding of Dynamic Medical CT Data Using Motion Compensated Wavelet Lifting with Denoised Prediction and Update 运动补偿小波提升与去噪预测与更新的医学CT动态数据无损编码
2019 Picture Coding Symposium (PCS) Pub Date : 2019-11-01 DOI: 10.1109/PCS48520.2019.8954530
Daniela Lanz, Franz Schilling, A. Kaup
{"title":"Scalable Lossless Coding of Dynamic Medical CT Data Using Motion Compensated Wavelet Lifting with Denoised Prediction and Update","authors":"Daniela Lanz, Franz Schilling, A. Kaup","doi":"10.1109/PCS48520.2019.8954530","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954530","url":null,"abstract":"Professional applications like telemedicine often require scalable lossless coding of sensitive data. 3-D subband coding has turned out to offer good compression results for dynamic CT data and additionally provides a scalable representation in terms of low- and highpass subbands. To improve the visual quality of the lowpass subband, motion compensation can be incorporated into the lifting structure, but leads to inferior compression results at the same time. Prior work has shown that a denoising filter in the update step can improve the compression ratio. In this paper, we present a new processing order of motion compensation and denoising in the update step and additionally introduce a second denoising filter in the prediction step. This allows for reducing the overall file size by up to 4.4%, while the visual quality of the lowpass subband is kept nearly constant.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129520787","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Bilateral Loop Filter in Combination with SAO 双侧环路滤波器与SAO的结合
2019 Picture Coding Symposium (PCS) Pub Date : 2019-11-01 DOI: 10.1109/PCS48520.2019.8954554
Jacob Ström, Per Wennersten, Jack Enhorn, Du Liu, K. Andersson, Rickard Sjöberg
{"title":"Bilateral Loop Filter in Combination with SAO","authors":"Jacob Ström, Per Wennersten, Jack Enhorn, Du Liu, K. Andersson, Rickard Sjöberg","doi":"10.1109/PCS48520.2019.8954554","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954554","url":null,"abstract":"This paper describes a bilateral filter that is being proposed as a coding tool for the Versatile Video Codec (VVC). The filter acts as a loop filter in parallel with the sample-adaptive offset (SAO) filter. Both the proposed filter and SAO act on the same input samples, each filter produces an offset, and these offsets are then added to the input sample to produce an output sample that, after clipping, goes to the next stage. The method has been implemented and tested according to the common test conditions in VVC test model version 5.0. For the all-intra configuration, we report a BD rate figure of -0.4% with an encoder run time increase of 6% and a decoder run time increase of 4%. For the random access configuration, the BD rate figure is -0.5% with an encoder run time increase of 2% and a decoder run time increase of 2%.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130023399","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Standard Designs for Cross Random Access Point Reference in Video Coding 视频编码中交叉随机接入点参考的标准设计
2019 Picture Coding Symposium (PCS) Pub Date : 2019-11-01 DOI: 10.1109/PCS48520.2019.8954520
Xiaoding Gao, Hualong Yu, Q. Yuan, Xiangyu Lin, Lu Yu
{"title":"Standard Designs for Cross Random Access Point Reference in Video Coding","authors":"Xiaoding Gao, Hualong Yu, Q. Yuan, Xiangyu Lin, Lu Yu","doi":"10.1109/PCS48520.2019.8954520","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954520","url":null,"abstract":"In videos like movies and TV shows, similar scenes usually appear alternately. The period between these similar scenes is so long that it even exceeds the length of a random access (RA) segment. This means that the temporal correlation among these similar scenes crosses random access point (RAP). Besides, in videos like surveillance videos, a similar scene usually lasts for a long time which is longer than a RA segment. In other words, the temporal correlation among this scene also crosses RAP. To exploit the cross-RAP temporal correlation for further improving coding performance and keep the random access functionality at the same time, library picture based cross random access point reference is proposed and adopted in AVS3, which is going to be introduced in this paper. Firstly, a library picture is used as a reference picture for similar RA segments to exploit the temporal correlation among these RA segments. Secondly, to support random access, decoder and system layer are both designed to ensure that the decoder can get the corresponding library picture when random access occurs to correctly decode RA segments. Experimental results show that this method can save 6.4% and 31.5% BD-rate on the AVS3 general sequences and special sequences.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124365210","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Fast Coding Unit Splitting Decisions for the Emergent AVS3 Standard 新兴AVS3标准的快速编码单元分割决策
2019 Picture Coding Symposium (PCS) Pub Date : 2019-11-01 DOI: 10.1109/PCS48520.2019.8954510
Meng Wang, Junru Li, Li Zhang, Kai Zhang, Hongbin Liu, Shiqi Wang, Siwei Ma
{"title":"Fast Coding Unit Splitting Decisions for the Emergent AVS3 Standard","authors":"Meng Wang, Junru Li, Li Zhang, Kai Zhang, Hongbin Liu, Shiqi Wang, Siwei Ma","doi":"10.1109/PCS48520.2019.8954510","DOIUrl":"https://doi.org/10.1109/PCS48520.2019.8954510","url":null,"abstract":"AVS3 adopts quad-tree (QT) with nested binary tree (BT) and extended quad-tree (EQT) partitioning, which shows promising compression performance when compared to the conventional QT partitioning in AVS2 and HEVC. However, the interleaved and recursive splitting manner significantly increases the computational complexity of the encoder, which may impede the real applications of AVS3. This paper proposes fast coding unit splitting decision methods for QT, BT and EQT partitioning. In particular, the Bayesian decision rule is employed with Skip states, and BT and EQT partitioning can be early terminated. Additionally, the average splitting depth that is produced by BT partitioning, is incorporated as the prior information for terminating the EQT and QT splittings. Moreover, we explore the splitting settings for CUs that are predicted with Skip mode, with the aim of eliminating unnecessary partition attempts. Experimental results show that the proposed fast algorithms are effective and provide a good trade-off between computational complexity and coding performance. In particular, 69% encoding time reduction is achieved with only 0.55% increase in terms of BD-Rate on average, which greatly benefits the practical implementations of the AVS3 in real applications. The proposed methods have been adopted into TAVS3 reference software.","PeriodicalId":237809,"journal":{"name":"2019 Picture Coding Symposium (PCS)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134010445","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信