2012 Visual Communications and Image Processing最新文献

筛选
英文 中文
A study on efficient compression of multi-focus images for dense Light-Field reconstruction 密集光场重建中多聚焦图像的有效压缩研究
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410759
Takashi Sakamoto, K. Kodama, T. Hamamoto
{"title":"A study on efficient compression of multi-focus images for dense Light-Field reconstruction","authors":"Takashi Sakamoto, K. Kodama, T. Hamamoto","doi":"10.1109/VCIP.2012.6410759","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410759","url":null,"abstract":"Light-Field enables us to observe scenes from free viewpoints. However, it generally consists of 4-D enormous data, that are not suitable for storing or transmitting without effective compression. 4-D Light-Field is very redundant because essentially it includes just 3-D scene information. Actually, although robust 3-D scene estimation such as depth recovery from Light-Field is not so easy, a method of reconstructing Light-Field directly from 3-D information composed of multi-focus images without any scene estimation is successfully derived. Previously, based on the method, Light-Field compression via synthesized multi-focus images as effective representation of 3-D scenes was proposed. In this paper, we study efficient compression of multi-focus images synthesized from dense Light-Field by using DWT instead of DCT-based compression in order to suppress degradation such as block noise. Quality of reconstructed Light-Field is evaluated by PSNR and SSIM for analyzing characteristics of residuals. Experimental results reveal that our method is much superior to Light-Field compression using disparity-compensation at low bit-rate.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133835820","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
A control-theoretic approach to rate adaptation for dynamic HTTP streaming 一种动态HTTP流速率自适应的控制理论方法
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410740
Chao Zhou, Xinggong Zhang, Longshe Huo, Zongming Guo
{"title":"A control-theoretic approach to rate adaptation for dynamic HTTP streaming","authors":"Chao Zhou, Xinggong Zhang, Longshe Huo, Zongming Guo","doi":"10.1109/VCIP.2012.6410740","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410740","url":null,"abstract":"Recently, dynamic adaptive HTTP streaming has been widely used for video content delivery over Internet. However, it is still a challenge how to switch video bitrate under time-varying bandwidth. In this paper, we propose a novel control-theoretic approach to adapt video segments in dynamic HTTP streaming. The rate control is based on a sink-buffer, which has an overflow-threshold and an underflow-threshold. The objective is to maximize the playback quality while keeping the receiver buffer from either overflow or underflow. Using control theory, we formulate this rate control scheme as a proportional (P) control system, which exists oscillations and steady-errors. Furthermore, we design a proportional derivative (PD) controller to improve its adaptation performance. The conditions for stability and settling time of the PD controller are also derived. Numerous experiment results demonstrate the effectiveness of our proposed PD control scheme for dynamic HTTP streaming.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132726925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 62
A novel no-reference image quality assessment metric based on statistical independence 一种基于统计独立性的无参考图像质量评价方法
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410790
Y. Chu, X. Mou, Wei Hong, Z. Ji
{"title":"A novel no-reference image quality assessment metric based on statistical independence","authors":"Y. Chu, X. Mou, Wei Hong, Z. Ji","doi":"10.1109/VCIP.2012.6410790","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410790","url":null,"abstract":"No-reference image quality assessment (NR IQA) has wide applicability to many problems. This paper focuses on the mechanism of divisive normalization transform (DNT) which simulates the behavior of visual cortex neurons to extract the independent components of natural images, analyzes the difference between the statistics of neighboring DNT coefficients of the images of a variety of distortion, and proposes a novel solution for NR IQA metric design. We demonstrate that measuring the statistical independence between neighboring DNT coefficients could provide features useful for quality assessment. The performance of the proposed method is quite satisfactory when it was tested on the popular LIVE, CSIQ and TID2008 databases. The experimental results are fairly competitive with the existing NR IQA metrics.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"62 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130070592","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Robust wheelchair pedestrian detection using sparse representation 基于稀疏表示的鲁棒轮椅行人检测
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410801
Po-Jui Huang, Duan-Yu Chen
{"title":"Robust wheelchair pedestrian detection using sparse representation","authors":"Po-Jui Huang, Duan-Yu Chen","doi":"10.1109/VCIP.2012.6410801","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410801","url":null,"abstract":"Detecting pedestrians with disability in surveillance videos is practical for the implementation of automated alert/assistance technology. This paper presents a novel approach for the dimensionality reduction which employs sparse representation to improve the generalization capability of a classifier. To characterize pedestrian with disability, we create directional maps by determining the dominant direction of motion in each local spatiotemporal region using 3D orientation filters, and then uses the maps in real-time surveillance settings to detect pre-defined types. Mathematically, the derived algorithm regards the input features as the dictionary in sparse representation, and selects the features that minimize the residual output error iteratively, thus the resulting features have a direct correspondence to the performance requirements of the given problem. Furthermore, the proposed algorithm can be regarded as a sparse classifier, which selects discriminative features and classifies the training data simultaneously. Experimental results obtained using the extensive dataset show the superior performance of our method and thus demonstrate its robustness with the novel sparse representation-based disabled pedestrian detector.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130312882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Background modeling using Local Binary Patterns Of Motion Vector 使用运动矢量局部二进制模式进行背景建模
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410784
Tingting Wang, Jiuzhen Liang, Xiaolong Wang, Shizheng Wang
{"title":"Background modeling using Local Binary Patterns Of Motion Vector","authors":"Tingting Wang, Jiuzhen Liang, Xiaolong Wang, Shizheng Wang","doi":"10.1109/VCIP.2012.6410784","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410784","url":null,"abstract":"Pixel-domain analysis methods are widely adopted in background modeling, some of which are not only concerned by academia but also coming into view of industry. However, as the increasing data volume of video, how to process and analysis videos in a fast and effective way has still been an intractable problem in practical applications. Under this circumstance, surveillance video analysis in the compressed domain is indeed of strategic importance from the angle of balancing visual perception and processing speed, especially in modeling background and segmenting moving objects. Therefore, a background modeling method in the compressed domain is proposed to quickly extract moving objects in this paper. Our main contributions are: 1) a method to calculate MVLBP features based on MV amplitude in the compressed domain is presented; 2) a background modeling and moving objects extraction method is designed in the compressed domain based on Local Binary Patterns of Motion Vector (MVLBP). Experimental results show that our approach gives a stable performance in a shorter time in H.264 compressed domain.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129846946","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A hierarchical mode decision scheme for fast implementation of spatially scalable video coding 一种用于快速实现空间可扩展视频编码的分层模式决策方案
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410746
Xin Lu, G. Martin
{"title":"A hierarchical mode decision scheme for fast implementation of spatially scalable video coding","authors":"Xin Lu, G. Martin","doi":"10.1109/VCIP.2012.6410746","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410746","url":null,"abstract":"In order to improve coding efficiency, a new inter-layer prediction mechanism is incorporated in the SVC extension of H.264/AVC. This utilizes information from the base layer to inform the process of coding the enhancement layer. However this increases the computational requirement. A fast mode decision algorithm that exploits the correlation of a macroblock in the enhancement layer and both the corresponding macroblock in the base layer and neighbouring macroblocks, is proposed. The algorithm also assesses the homogeneity of the picture content and uses the mode information of the base layer to make faster mode selections in the enhancement layer. The fact that larger partition sizes are more suitable for homogeneous regions, and vice versa, is also exploited. Empirical evaluation of the proposed algorithm shows that, for similar rate distortion performance, encoding time is reduced by up to 84% compared with the JSVM9.18 software implementation.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129205046","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Fast and reliable noise estimation algorithm based on statistical hypothesis tests 基于统计假设检验的快速可靠的噪声估计算法
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410754
Ping Jiang, Jianzhou Zhang
{"title":"Fast and reliable noise estimation algorithm based on statistical hypothesis tests","authors":"Ping Jiang, Jianzhou Zhang","doi":"10.1109/VCIP.2012.6410754","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410754","url":null,"abstract":"Image noise estimation is a very important topic in digital image processing. This paper presents a fast and reliable noise estimation algorithm for additive white Gaussian noise (WGN). The proposed algorithm provides a way to measure the degree of image feature based on statistical hypothesis tests (SHT). Firstly, the proposed algorithm distinguishes homogeneous blocks and non-homogeneous blocks by the degree of image feature, and then sets the minimal variance of these homogeneous blocks as a reference variance. Secondly, the proposed algorithm finds more homogeneous blocks whose variances are similar to the reference variance and which are not non-homogeneous blocks. Lastly, the noise variance is estimated from these homogeneous blocks by a weighted averaging process according to the degree of image feature. Experiments show that the proposed algorithm performs well and reliably for different types of images over a large range of noise levels.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127582320","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
User-adaptive mobile video streaming 用户自适应移动视频流
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410862
Y. Reznik, E. Asbun, Z. Chen, Yan Ye, E. Zeira, R. Vanam, Zheng Yuan, Gregory Sternberg, A. Zeira, Naresh Soni
{"title":"User-adaptive mobile video streaming","authors":"Y. Reznik, E. Asbun, Z. Chen, Yan Ye, E. Zeira, R. Vanam, Zheng Yuan, Gregory Sternberg, A. Zeira, Naresh Soni","doi":"10.1109/VCIP.2012.6410862","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410862","url":null,"abstract":"Summary form only given. We describe the design of a mobile streaming system, which optimizes video delivery based on dynamic analysis of user behavior and viewing conditions, including user proximity, viewing angle, and ambient illuminance.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121411887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
The hierarchical signal dependent transform: Creating orthonormal basis that match local signal characteristics 层次信号相关变换:创建匹配局部信号特征的标准正交基
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410820
Vanessa Testoni, M. H. M. Costa, Dinei Fiorendo
{"title":"The hierarchical signal dependent transform: Creating orthonormal basis that match local signal characteristics","authors":"Vanessa Testoni, M. H. M. Costa, Dinei Fiorendo","doi":"10.1109/VCIP.2012.6410820","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410820","url":null,"abstract":"Hierarchical transforms are widely used in image and video coding to produce multilevel decomposition of signals. After applying these transforms, same level signals are typically uncorrelated. However, there is often still significant cross level information. Traditionally, this cross-level information is exploited at the entropy coding step, but not at the transform step. The main contribution of this work is an energy compaction technique/transform that can also exploit these cross-resolution-level structural similarities. The core idea of the technique is to include in the hierarchical transform a number of adaptive basis functions derived from the lower resolution of the signal. A full image codec was developed in order to measure the performance of the new transform. Results are presented in terms of transform coding gain, energy concentration and distortion versus rate curves compared with standard JPEG, JPEG 2000 and JPEG XR.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121259126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Layered compression for high dynamic range depth 分层压缩的高动态范围深度
2012 Visual Communications and Image Processing Pub Date : 2012-11-01 DOI: 10.1109/VCIP.2012.6410786
Dan Miao, Jingjing Fu, Yan Lu, Shipeng Li, Chang Wen Chen
{"title":"Layered compression for high dynamic range depth","authors":"Dan Miao, Jingjing Fu, Yan Lu, Shipeng Li, Chang Wen Chen","doi":"10.1109/VCIP.2012.6410786","DOIUrl":"https://doi.org/10.1109/VCIP.2012.6410786","url":null,"abstract":"With the rapid development of depth data acquisition technology, the high precision depth becomes much easier to access in real-time by depth sensors, and the generated high dynamic range (HDR) depth is widely adopted to benefit the depth assistant applications. Accordingly, the HDR depth compression becomes essential for the efficient depth storage and transmission. In this paper, we introduce a layered compression framework for HDR depth to achieve efficient and low-complexity depth compression. To leverage the state-of-art 8-bit image/video encoders, the HDR depth is partitioned into two layers: most significant bit (MSB) layer and least significant bit (LSB) layer. For MSB layer, an error controllable pixel domain encoding scheme is proposed to guarantee the compatibility for existing 8-bit codec by controlling quantization errors added back to LSB layer. Meanwhile, the efficient major color extraction and adaptive quantization enhance the coding performance of MSB layer. For LSB layer, the layer data with limited dynamic range is compressed by normal 8-bit image/video based encoding scheme. The experimental results demonstrate that our coding scheme can achieve real-time depth compression with the satisfactory reconstruction quality. The encoding time is less than 31ms/frame and the decoding time is around 20ms/frame in average. Our compression scheme can be easily integrated into the real-time depth transmission system.","PeriodicalId":103073,"journal":{"name":"2012 Visual Communications and Image Processing","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121540710","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信