2012 Visual Communications and Image Processing最新文献

筛选
英文 中文
Implementation of HEVC decoder on x86 processors with SIMD optimization 基于SIMD优化的x86处理器上HEVC解码器的实现
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410845
Leju Yan, Y. Duan, Jun Sun, Zongming Guo
{"title":"Implementation of HEVC decoder on x86 processors with SIMD optimization","authors":"Leju Yan, Y. Duan, Jun Sun, Zongming Guo","doi":"10.1109/VCIP.2012.6410845","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410845","url":null,"abstract":"High Efficient Video Coding (HEVC) is the next generation video coding standard in progress. Based on the traditional hybrid coding framework, HEVC implements enhanced tools to improve compression efficiency at the cost of far more computational payload than the capacity of real-time video applications. In this paper, we focus on the software implementation of a real-time HEVC decoder over modern Intel x86 processors. First, we identify the most time-consuming modules of HM 4.0 decoder, represented by motion compensation, adaptive loopfilter, deblocking filter and integer transform. Then the single-execution-multiple-data (SIMD) methods are proposed to optimize the computational performance of these modules. Experimental results show that the optimized decoder is more than 4 times faster than the HM 4.0 decoder, with decoding speed of over 40 frames per second for 1920×1080 resolution videos on Intel i5-2400 processor.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"104 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122787913","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
Low-complexity and high-efficiency background modeling for surveillance video coding 低复杂度、高效率的监控视频编码背景建模
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410796
Xianguo Zhang, Yonghong Tian, Tiejun Huang, Wen Gao
{"title":"Low-complexity and high-efficiency background modeling for surveillance video coding","authors":"Xianguo Zhang, Yonghong Tian, Tiejun Huang, Wen Gao","doi":"10.1109/VCIP.2012.6410796","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410796","url":null,"abstract":"Recently, background modeling (shortly BgModeling) plays a more and more important role in high-efficiency surveillance video coding. Meanwhile, many practical video coding applications also present some specific requirements for BgModeling, such as the low memory cost and low computational complexity. However, existing BgModeling methods are mostly designed for video content analysis such as object detection. Thus they may be not directly applicable for video coding. In this paper, we firstly present an analysis for the features of BgModeling in surveillance video coding and make a comparison of the performances of existing BgModeling methods. Then we propose a segment-and-weight based running average (SWRA) method for surveillance video coding. SWRA firstly divides pixels at each position in the training frames into several temporal segments, and then calculate their corresponding mean values and weights. After that, a running and weighted average procedure is used to reduce the influence of foreground pixels and finally obtain the modeling results. Experimental results show that, the SWRA-based encoder achieves the best performance over several state-of-the-art methods, with much less cost of memory and modeling time.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115222991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
Improving spatio-temporal feature extraction techniques and their applications in action classification 改进时空特征提取技术及其在动作分类中的应用
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410811
Maral Mesmakhosroshahi, Joohee Kim
{"title":"Improving spatio-temporal feature extraction techniques and their applications in action classification","authors":"Maral Mesmakhosroshahi, Joohee Kim","doi":"10.1109/VCIP.2012.6410811","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410811","url":null,"abstract":"Space-time feature extraction is a recent and popular method used for action recognition. This paper presents a new algorithm to improve the robustness of spatio-temporal feature extraction techniques against the illumination and scale variations. Most of the interest point detectors are sensitive to illumination variations that may cause serious problems in action recognition by finding wrong keypoints. A method is proposed to make the 3-D Harris corner detector robust to illumination changes. Illumination invariance is achieved by applying a contrast stretching function to the video to find the proper intensity level for each pixel. A non-uniform binning method is also proposed to make the 3-D extension of the well-known SIFT descriptor more reliable and robust to scale changes by forming orientation histograms which concentrate on the regions near the interest points. Bag of features technique is used for classifying actions provided by the KTH dataset and the results demonstrate that our proposed method outperforms the original 3-D corner detector and SIFT descriptor.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126877615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Gradient-based fast decision for intra prediction in HEVC 基于梯度的HEVC内预测快速决策
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410739
Yongfei Zhang, Zhe Li, Bo Li
{"title":"Gradient-based fast decision for intra prediction in HEVC","authors":"Yongfei Zhang, Zhe Li, Bo Li","doi":"10.1109/VCIP.2012.6410739","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410739","url":null,"abstract":"As the next generation standard of video coding, the High Efficiency Video Coding(HEVC) achieves significantly better coding efficiency than all existing video coding standards, which is however at the cost of a much higher computation complexity. To address this issue, this paper presents a gradient-based fast decision algorithm for intra prediction in HEVC. More specifically, the intra prediction in HEVC is divided into two stages: prediction unit(PU) size decision and mode decision. At the PU size decision process, four orientation features are extracted from the coding unit by the intensity gradient filters to decide the texture complexity and texture direction of the coding unit, and then the texture direction is used to exclude impossible prediction modes at the mode decision process. Compared to HEVC reference software, the proposed algorithm saves around 56.7% of the encoding time in intra high efficiency setting and up to 70.86% in intra low complexity setting with slight performance degradation.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116088123","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 79
Spatio-temporal ssim index for video quality assessment 用于视频质量评价的时空ssim指数
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410779
Yue Wang, Tingting Jiang, Siwei Ma, Wen Gao
{"title":"Spatio-temporal ssim index for video quality assessment","authors":"Yue Wang, Tingting Jiang, Siwei Ma, Wen Gao","doi":"10.1109/VCIP.2012.6410779","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410779","url":null,"abstract":"An ideal objective metric for video quality assessment (VQA) should achieve consistency between video distortion prediction and psychological perception of human visual system (HVS), and is important in many video processing applications. In general, both spatial distortion and temporal distortion should be carefully considered in the designing of VQA metrics. In this paper, we propose a novel spatio-temporal structural information based video quality metric. Motivated by the fact that pixels in natural videos are highly structured in both spatial domain and temporal domain, we propose to perform structural similarity evaluation in x-y, x-t and y-t dimensions respectively and pooled them adaptively based on local spatio-temporal activities. Experimental results on LIVE database show that such a conceptually simple and computationally efficient algorithm is competitive with state-of-the-art VQA metrics, and is very robust to various types of video distortions.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121587994","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Region-Based perceptual quality regulable bit allocation and rate control for video coding applications 基于区域的感知质量可调比特分配和视频编码速率控制
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410836
Guan-Lin Wu, Yu-Jie Fu, Shao-Yi Chien
{"title":"Region-Based perceptual quality regulable bit allocation and rate control for video coding applications","authors":"Guan-Lin Wu, Yu-Jie Fu, Shao-Yi Chien","doi":"10.1109/VCIP.2012.6410836","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410836","url":null,"abstract":"In this paper, a perceptual quality regulable H.264 video encoder system has been developed. We use structure similarity index as the quality metric for distortion-quantization modeling and develop a bit allocation and rate control scheme for enhancing regional perceptual quality. Exploiting the relationship between the reconstructed macroblock and its best predicted macroblock from mode decision, a novel quantization parameter prediction method is built and used to regulate the video quality of the processing macroblock according to a target perceptual quality. Experimental results show that the model can achieve high accurate. Compared to JM reference software with macroblock layer rate control, the proposed encoding system can effectively enhance perceptual quality for target video regions.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"62 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116952261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Improvement of normality and orthogonality in HEVC transform bases HEVC变换基正正交性的改进
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410730
Yasuko Sugito, A. Ichigaya, S. Sakaida
{"title":"Improvement of normality and orthogonality in HEVC transform bases","authors":"Yasuko Sugito, A. Ichigaya, S. Sakaida","doi":"10.1109/VCIP.2012.6410730","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410730","url":null,"abstract":"The present paper provides transform bases with improved normality and orthogonality properties based on the integer DCT of high-efficiency video coding (HEVC). The proposed transform bases improve the normality and orthogonality properties compared to the HEVC transform bases. Coding and re-encoding experiments were conducted using HEVC test model (HM) version 6.0 and the proposed method. The experiments were conducted under various bitrate ranges. Under the high-bit-rate condition, the proposed method exhibited coding gains compared to HM 6.0 without increasing the encoding/decoding time.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115058844","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integrating stereoscopic image transcoding with retargeting for mobile streaming 整合立体图像转码与移动流的重定向
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410756
Yanwei Liu, S. Ci, Jinxia Liu, Yun Ye, Y. Qi
{"title":"Integrating stereoscopic image transcoding with retargeting for mobile streaming","authors":"Yanwei Liu, S. Ci, Jinxia Liu, Yun Ye, Y. Qi","doi":"10.1109/VCIP.2012.6410756","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410756","url":null,"abstract":"With the progresses in 3D imaging technologies, mobile stereoscopic images are gradually popular since it can provide the anytime and anywhere 3D viewing effects. Due to the heterogeneous transmission environments and various screen sizes of mobile devices, the stereoscopic image transcoding and retargeting are two completely independent adaptation techniques for mobile 3D image streaming. These two processing operations both deal with the disparity estimation. In this work, we propose to integrate stereoscopic 3D image transcoding with retargeting for mobile streaming. The stereoscopic image transcoding and retargeting are coupled through a straightforward link that provides the pixel-based disparity map from the retargeting stage to guide the transcoding for saving some computations while keeping efficient bit allocation. Our experimental results clearly demonstrate the advantage of coupling the transcoding with retargeting in terms of faster transcoding and saliency-based bit allocation for low bit-rate mobile streaming.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115180355","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Error resilient video coding for systems with delayed feedback 延迟反馈系统的抗错误性视频编码
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410800
R. Vanam, Z. Chen, Y. Reznik
{"title":"Error resilient video coding for systems with delayed feedback","authors":"R. Vanam, Z. Chen, Y. Reznik","doi":"10.1109/VCIP.2012.6410800","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410800","url":null,"abstract":"In systems employing feedback-based error resilience, error propagation can significantly degrade visual quality when feedback delay is in the order of a few seconds. We propose a coding structure based on multiple description coding that mitigates error propagation during feedback delay, and uses feedback to adapt its coding structure to effectively limit error propagation. We demonstrate the effectiveness of our approach at different error rates when compared to conventional coding schemes that use feedback.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123634535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A low-complexity screen compression scheme 一个低复杂度的屏幕压缩方案
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410780
Zhaotai Pan, Huifeng Shen, Yan Lu, Nenghai Yu, Shipeng Li
{"title":"A low-complexity screen compression scheme","authors":"Zhaotai Pan, Huifeng Shen, Yan Lu, Nenghai Yu, Shipeng Li","doi":"10.1109/VCIP.2012.6410780","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410780","url":null,"abstract":"This paper presents a block-based low-complexity screen compression scheme. In this scheme, the input screen is split into non-overlapping blocks which are classified as pictorial blocks and textual blocks. We design a low-complexity yet efficient algorithm to compress the textual blocks. We use base colors plus escape pixels to represent and quantize the text pixels, and such quantization mechanism not only achieves high compression efficiency but also keeps low encoding/decoding complexity. We also propose the two-direction predictive index coding and hierarchical pattern coding technologies to utilize the local spatial correlation and the global pattern correlation for text pixels. In addition, to utilize the correlation between the luminance and chrominance channels, we propose a joint-channel index coding method to further improve the compression efficiency. The compression efficiency and complexity of the proposed method is evaluated against the popular image codecs JPEG/JPEG200 and PNG, a recent published screen compression scheme HJPC as well as the popular video codec H.264.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127871892","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信