28th Picture Coding Symposium最新文献

筛选
英文 中文
Edge-adaptive transforms for efficient depth map coding 边缘自适应变换用于有效的深度图编码
28th Picture Coding Symposium Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702565
Godwin Shen, Woo-Shik Kim, S. K. Narang, Antonio Ortega, Jaejoon Lee, H. Wey
{"title":"Edge-adaptive transforms for efficient depth map coding","authors":"Godwin Shen, Woo-Shik Kim, S. K. Narang, Antonio Ortega, Jaejoon Lee, H. Wey","doi":"10.1109/PCS.2010.5702565","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702565","url":null,"abstract":"In this work a new set of edge-adaptive transforms (EATs) is presented as an alternative to the standard DCTs used in image and video coding applications. These transforms avoid filtering across edges in each image block, thus, they avoid creating large high frequency coefficients. These transforms are then combined with the DCT in H.264/AVC and a transform mode selection algorithm is used to choose between DCT and EAT in an RD-optimized manner. These transforms are applied to coding depth maps used for view synthesis in a multi-view video coding system, and provides up to 29% bit rate reduction for a fixed quality in the synthesized views.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126663786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 178
Novel video coding paradigm with reduction/restoration processes 新的视频编码范式与减少/恢复过程
28th Picture Coding Symposium Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702537
T. Misu, Y. Matsuo, S. Sakaida, Y. Shishikui, E. Nakasu
{"title":"Novel video coding paradigm with reduction/restoration processes","authors":"T. Misu, Y. Matsuo, S. Sakaida, Y. Shishikui, E. Nakasu","doi":"10.1109/PCS.2010.5702537","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702537","url":null,"abstract":"To optimally design distortions in lossy video coding, we propose the use of a novel coding paradigm with adaptive nonlinear transforms as pre/post-processors of a conventional video codec. The preprocessor decimates less important pixels based on an image analysis. A conventional video encoder such as MPEG-4 AVC/H.264 further eliminates the redundancy of the decimated images. On the decoder side, the postprocessor restores small decoded images of the conventional decoder to the original resolution using an inverse mapping including a super-resolution technique that uses a priori knowledge on the decimation in the preprocessing. Results of the experiments showed the proposed coding scheme poses distortion that has a more straightforward appearance than that of a directly encoded/decoded image by a sole conventional H.264 codec.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126754159","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
A novel inloop filter for video-compression based on temporal pixel trajectories 一种基于时间像素轨迹的视频压缩内环滤波器
28th Picture Coding Symposium Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702551
Marko Esche, A. Krutz, A. Glantz, T. Sikora
{"title":"A novel inloop filter for video-compression based on temporal pixel trajectories","authors":"Marko Esche, A. Krutz, A. Glantz, T. Sikora","doi":"10.1109/PCS.2010.5702551","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702551","url":null,"abstract":"The objective of this work is to investigate the performance of a new inloop filter for video compression, which uses temporal rather than spatial information to improve the quality of reference frames used for prediction. The new filter has been integrated into the H.264/AVC baseline encoder and tested on a wide range of sequences. Experimental results show that the filter achieves a bit rate reduction of up to 12% and more than 4% on average without increasing the complexity of either encoder or decoder significantly.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"148 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115435870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
3D space representation using epipolar plane depth image 利用极平面深度图像进行三维空间表示
28th Picture Coding Symposium Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702472
Takashi Ishibashi, T. Yendo, M. P. Tehrani, T. Fujii, M. Tanimoto
{"title":"3D space representation using epipolar plane depth image","authors":"Takashi Ishibashi, T. Yendo, M. P. Tehrani, T. Fujii, M. Tanimoto","doi":"10.1109/PCS.2010.5702472","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702472","url":null,"abstract":"We propose a novel 3D space representation for multi-view video, using epipolar plane depth images (EPDI). Multi-view video plus depth (MVD) is used as common data format for FTV(Free-viewpoint TV), which enables synthesizing virtual view images. Due to large amount of data and complexity of the multi-view video coding (MVC), compression of MVD is a challenging issue. We address this problem and propose a new representation that is constructed from MVD using rayspace. MVD is converted into image and depth ray-spaces. The proposed representation is obtained by converting each of ray-spaces into a global depth map and a texture map using EPDI. Experiments demonstrate the analysis of this representation, and its efficiency.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125948160","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Towards efficient intra prediction based on image inpainting methods 基于图像绘制方法的高效图像内预测
28th Picture Coding Symposium Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702539
D. Doshkov, P. Ndjiki-Nya, H. Lakshman, Martin Köppel, T. Wiegand
{"title":"Towards efficient intra prediction based on image inpainting methods","authors":"D. Doshkov, P. Ndjiki-Nya, H. Lakshman, Martin Köppel, T. Wiegand","doi":"10.1109/PCS.2010.5702539","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702539","url":null,"abstract":"In this paper, novel intra prediction methods based on image inpainting approaches are proposed. The H.264/AVC intra prediction modes are not well suited for processing complex textures at low bit rates. Our algorithm utilizes an efficient combination of partial differential equations (PDEs) and patch-based texture synthesis in addition to the standard directional predictors. Bit rate savings up to 3.5% compared to that of the H.264/AVC standard are shown.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121124527","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
Theoretical analysis of trend vanishing moments for directional orthogonal transforms 方向正交变换趋势消失矩的理论分析
28th Picture Coding Symposium Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702440
S. Muramatsu, Dandan Han, Tomoya Kobayashi, H. Kikuchi
{"title":"Theoretical analysis of trend vanishing moments for directional orthogonal transforms","authors":"S. Muramatsu, Dandan Han, Tomoya Kobayashi, H. Kikuchi","doi":"10.1109/PCS.2010.5702440","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702440","url":null,"abstract":"This work contributes to investigate theoretical properties of the trend vanishing moment (TVM) which the authors have defined in a previous work and applied to the directional design of 2-D nonsep-arable GenLOTs. The TVM condition can be regarded as a natural extention of 1-D VM to 2-D one and alternative of the conventional directional vanishing moment (DVM). Firstly, the definition of TVM is given to clarify what it is, and then the mathematical meaning is discussed. While the conventional DVM condition requires for the moments to vanish along lines in the frequency domain and restricts the direction to a rational factor, the TVM condition imposes the moments only point-wisely and the direction can be steered flexibly. Some significant properties of TVMs are also shown and a simulation result of zonal coding for an artifical picture is given to verify the trend surface annihilation property.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125992935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A novel coding scheme for intra pictures of H.264/AVC 一种新的H.264/AVC帧内图像编码方案
28th Picture Coding Symposium Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702579
Jin Young Lee, Jaejoon Lee, H. Wey, Du-sik Park
{"title":"A novel coding scheme for intra pictures of H.264/AVC","authors":"Jin Young Lee, Jaejoon Lee, H. Wey, Du-sik Park","doi":"10.1109/PCS.2010.5702579","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702579","url":null,"abstract":"A novel intra coding scheme is proposed to improve coding performance in intra pictures of H.264/AVC. The proposed method generates two sub-images, which are defined as a sampled image and a prediction error image in this paper, from an original image, and then encodes them separately. Especially, in an intra prediction process of encoding, the sampled image employs the original intra prediction modes, while the prediction error image uses newly defined four intra prediction modes. Experimental results demonstrate that the proposed method achieves significantly higher intra coding performance and reduces encoding complexity with the smaller number of a rate-distortion (RD) cost calculation process, as compared with the original intra coding method of H.264/AVC.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134254449","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Fractional-sample motion compensation using generalized interpolation 基于广义插值的分数样本运动补偿
28th Picture Coding Symposium Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702555
H. Lakshman, B. Bross, H. Schwarz, T. Wiegand
{"title":"Fractional-sample motion compensation using generalized interpolation","authors":"H. Lakshman, B. Bross, H. Schwarz, T. Wiegand","doi":"10.1109/PCS.2010.5702555","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702555","url":null,"abstract":"Typical interpolation methods in video coding perform filtering of reference picture samples using FIR filters for motion-compensated prediction. This process can be viewed as a signal decomposition using basis functions which are restricted by the interpolating constraint. Using the concept of generalized interpolation provides a greater degree of freedom for selecting basis functions. We implemented generalized interpolation using a combination of IIR and FIR filters. The complexity of the proposed scheme is comparable to that of an 8-tap FIR filter. Bit rate savings up to 20% compared to the H.264/AVC 6-tap filter are shown.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133091209","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
A study on memorability and shoulder-surfing robustness of graphical password using DWT-based image blending 基于dwt图像混合的图形密码记忆性和肩部冲浪鲁棒性研究
28th Picture Coding Symposium Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702441
T. Miyachi, Keita Takahashi, Madoka Hasegawa, Yuichi Tanaka, Shigeo Kato
{"title":"A study on memorability and shoulder-surfing robustness of graphical password using DWT-based image blending","authors":"T. Miyachi, Keita Takahashi, Madoka Hasegawa, Yuichi Tanaka, Shigeo Kato","doi":"10.1109/PCS.2010.5702441","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702441","url":null,"abstract":"Graphical passwords are an authentication method that uses pictures as passwords instead of using alphanumeric characters. We propose a graphical password method which is difficult to steal original pass-image by using characteristics of human vision system. In our method, we combine low frequency components of a decoy picture with high frequency components of a pass-image. It is easy for legitimate users to recognize the pass-image in the blended image. On the other hand, this task is difficult for attackers. We used discrete wavelet transform (DWT) to blend a decoy image and a pass-image. User studies are conducted to evaluate memorability and shoulder-surfing robustness of this method. We also compared our method with other existing methods in terms of the authentication time and the success ratio by the user test. The results show that our method is more usable and secure against shoulder-surfing.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133893504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A background model based method for transcoding surveillance videos captured by stationary camera 一种基于背景模型的静止摄像机监控视频转码方法
28th Picture Coding Symposium Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702583
Xianguo Zhang, Luhong Liang, Qian Huang, Tiejun Huang, Wen Gao
{"title":"A background model based method for transcoding surveillance videos captured by stationary camera","authors":"Xianguo Zhang, Luhong Liang, Qian Huang, Tiejun Huang, Wen Gao","doi":"10.1109/PCS.2010.5702583","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702583","url":null,"abstract":"Real-world video surveillance applications require storing videos without neglecting any part of scenarios for weeks or months. To reduce the storage cost, the high bit-rate videos from cameras should be transcoded into a more efficient compressed format with as little quality loss as possible. In this paper, we propose a background model based method to improve the transcoding efficiency for surveillance videos captured by stationary cameras, and objectively measure it. The background model is trained by pre-decoded I frames, and then used to transcode the source stream. Following this method, an H.264/AVC based transcoder employing the background model as long-term reference frame and a difference frame coding based transcoder are implemented and evaluated. Experimental results show that both trancoders save nearly half the used bits while maintaining quality compared with the full-decoding-full-encoding method, and the latter one has slightly better performance.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129519781","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信