2014 19th International Conference on Digital Signal Processing最新文献

筛选
英文 中文
A low complexity and high performance interpolation filter for MPEG IVC 一个低复杂度和高性能的mpegivc插值滤波器
2014 19th International Conference on Digital Signal Processing Pub Date : 2014-09-18 DOI: 10.1109/ICDSP.2014.6900817
Hao Lv, Ronggang Wang, Yangang Cai, Huizhu Jia, Xiaodong Xie, Wen Gao
{"title":"A low complexity and high performance interpolation filter for MPEG IVC","authors":"Hao Lv, Ronggang Wang, Yangang Cai, Huizhu Jia, Xiaodong Xie, Wen Gao","doi":"10.1109/ICDSP.2014.6900817","DOIUrl":"https://doi.org/10.1109/ICDSP.2014.6900817","url":null,"abstract":"Fractional-pel motion compensation is widely adopted in the modern video coding standards such as H.264/AVC, AVS, and HEVC etc. The interpolation filter is a critical factor that influences the coding efficiency. In this paper, a generation algorithm of interpolation filter coefficients is utilized. Based on the coefficients generation algorithm, three different tap filters, namely 6 tap, 8 tap, 10tap, are tested. A combination of 6 tap and 8 tap interpolation filters is proposed and proved to be optimal scheme considering both performance and computational complexity. And it is beneficial to make software optimization more effective, especially when SIMD-like (Single Instruction Multiple Data) operation is used. Experiments show that the average BD-rate gains on luma Y, chroma U and V are 8.01%, 5.08% and 4.98% for CS1 (Constraint set 1), and 9.21%, 7.53% and 7.63% for CS2 (Constraint set 2) in MPEG IVC reference software ITM5.0, compared with the traditional IVC interpolation method. The coding efficiency gains are significant for some video sequences and can reach up to 28.7%. With the merits of high performance and low complexity, our proposed method is formally adopted by MPEG IVC.","PeriodicalId":301856,"journal":{"name":"2014 19th International Conference on Digital Signal Processing","volume":"154 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114642953","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Green noise video halftoning 绿噪视频半调
2014 19th International Conference on Digital Signal Processing Pub Date : 2014-09-18 DOI: 10.1109/ICDSP.2014.6900807
Yik-Hing Fung, Y. Chan
{"title":"Green noise video halftoning","authors":"Yik-Hing Fung, Y. Chan","doi":"10.1109/ICDSP.2014.6900807","DOIUrl":"https://doi.org/10.1109/ICDSP.2014.6900807","url":null,"abstract":"Video halftoning is a technology used to render a video onto a display device that can only display limited number of levels. Conventional video halftoning algorithms produce blue noise video halftones which are prone to flickering. Dedicated deflickering processes are hence required to reduce flickering. These processes share a common approach in which pixels are artificially made stable subject to some quality constraints. Due to the difficulty to control the extent of stability, artifacts caused by overstability such as dirty window effect, subtle motion and residual shadow are easily found in video halftones. In this paper, we suggest producing green noise video halftones instead of blue noise video halftones. By doing so, we can effectively reduce flickering and eliminate artifacts caused by overstability from the root simultaneously.","PeriodicalId":301856,"journal":{"name":"2014 19th International Conference on Digital Signal Processing","volume":"115 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115125305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Novel CAVLC design for secondary SP-frame 新颖的二次sp车架CAVLC设计
2014 19th International Conference on Digital Signal Processing Pub Date : 2014-09-18 DOI: 10.1109/ICDSP.2014.6900833
Hongbin Zhang, C. Fu, W. Su, Yui-Lam Chan
{"title":"Novel CAVLC design for secondary SP-frame","authors":"Hongbin Zhang, C. Fu, W. Su, Yui-Lam Chan","doi":"10.1109/ICDSP.2014.6900833","DOIUrl":"https://doi.org/10.1109/ICDSP.2014.6900833","url":null,"abstract":"The quantized DCT coefficients of secondary SP-frame are different from conventional P-frame statistically. Therefore, the context-based adaptive variable length coder (CAVLC) designed in H.264 is not suitable for secondary SP-frame. A novel entropy coding approach based on CAVLC is designed for secondary SP-frame according to the special characteristic. The approach selects better VLC table for the level coding of nonzero coefficients by modeling the data fluctuation in each block. Moreover, tables of coding nonzero coefficients and trailingones are also redesigned. Simulation results show that the proposed approach can save 2.04% bits in average compared with H.264/AVC.","PeriodicalId":301856,"journal":{"name":"2014 19th International Conference on Digital Signal Processing","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115518736","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multi-resolution PR NMDFBs for programmable variable bandwidth filter in wideband digital transceivers 用于宽带数字收发器可编程可变带宽滤波器的多分辨率PR nmdfb
2014 19th International Conference on Digital Signal Processing Pub Date : 2014-09-18 DOI: 10.1109/ICDSP.2014.6900809
F. Harris, Behrokh Farzad, E. Venosa, Xiaofei Chen
{"title":"Multi-resolution PR NMDFBs for programmable variable bandwidth filter in wideband digital transceivers","authors":"F. Harris, Behrokh Farzad, E. Venosa, Xiaofei Chen","doi":"10.1109/ICDSP.2014.6900809","DOIUrl":"https://doi.org/10.1109/ICDSP.2014.6900809","url":null,"abstract":"This paper describes a novel application of perfect reconstruction (PR) non-maximally decimated filter banks (NMDFBs). PR non-maximally decimated analysis/synthesis chains are used for performing resampling, channelization and filtering in very wideband software defined radios. Digital wideband filtering is currently limited by the hardware clock speed which constrains the maximum signal bandwidth that can be processed digitally. Polyphase channelizers decrease the sample rate of the input signals and parallelize the wideband filtering process at cost of a single filter. In this paper we show how to assemble multiple-tier PR analysis/synthesis channelizer chains for implementing, in a very efficient way, variable bandwidth digital filters and multi-resolution systems operating on very wideband signals. When an internal tier analysis channelizer is applied to a single output channel of a first external tier analysis channelizer it decomposes this channel in multiple narrower subchannels and, by applying appropriate complex processing elements to those sub-channels we can modify the designed filter bandwidth as desired.","PeriodicalId":301856,"journal":{"name":"2014 19th International Conference on Digital Signal Processing","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123458030","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Corticomuscular coherence analysis on the static and dynamic tasks of hand movement 手部运动静态和动态任务的皮质肌肉一致性分析
2014 19th International Conference on Digital Signal Processing Pub Date : 2014-09-18 DOI: 10.1109/ICDSP.2014.6900757
Anshuang Fu, Rui Xu, Feng He, Hongzhi Qi, Lixin Zhang, Dong Ming, Yanru Bai, Zhiguo Zhang
{"title":"Corticomuscular coherence analysis on the static and dynamic tasks of hand movement","authors":"Anshuang Fu, Rui Xu, Feng He, Hongzhi Qi, Lixin Zhang, Dong Ming, Yanru Bai, Zhiguo Zhang","doi":"10.1109/ICDSP.2014.6900757","DOIUrl":"https://doi.org/10.1109/ICDSP.2014.6900757","url":null,"abstract":"The synchronization between cortical motor and muscular activity can be revealed by corticomuscular coherence (CMC). This paper designed two neuromuscular activity paradigms of hand movement, i.e. static gripping task and dynamic finger moving task. The electroencephalography (EEG) from C3 and C4 channels and the surface electromyography (sEMG) from the flexor digitorum superficialis were collected simultaneously from 4 male and 4 female right-handed healthy young subjects. For the static griping task, CMCs during low-level forces under 4%, 8%, and 16% MVC (Maximal Voluntary Contraction) were investigated by using magnitude squared coherence calculated from EEGs and sEMGs. For the dynamic finger moving task, the time-frequency domain analysis was used to process dynamic data of temporary action in a period of 2 seconds and get the latency of the maximum CMC. The results of this study indicated that the force increasing within the low-level range in static task is associated with the enhanced CMC. The maximum amplitude of CMC occurred about 0.3-0.5s after the onset of hand movement. Subjects showed significant CMC performance both in static and dynamic task of hand movement.","PeriodicalId":301856,"journal":{"name":"2014 19th International Conference on Digital Signal Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128803725","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Signal classification based on block-sparse tensor representation 基于块稀疏张量表示的信号分类
2014 19th International Conference on Digital Signal Processing Pub Date : 2014-09-18 DOI: 10.1109/ICDSP.2014.6900687
S. Zubair, Wenwu Wang
{"title":"Signal classification based on block-sparse tensor representation","authors":"S. Zubair, Wenwu Wang","doi":"10.1109/ICDSP.2014.6900687","DOIUrl":"https://doi.org/10.1109/ICDSP.2014.6900687","url":null,"abstract":"Block sparsity was employed recently in vector/matrix based sparse representations to improve their performance in signal classification. It is known that tensor based representation has potential advantages over vector/matrix based representation in retaining the spatial distributions within the data. In this paper, we extend the concept of block sparsity for tensor representation, and develop a new algorithm for obtaining sparse tensor representations with block structure. We show how the proposed algorithm can be used for signal classification. Experiments on face recognition are provided to demonstrate the performance of the proposed algorithm, as compared with several sparse representation based classification algorithms.","PeriodicalId":301856,"journal":{"name":"2014 19th International Conference on Digital Signal Processing","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129932460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Parallelized feature extraction and acoustic model training 并行特征提取和声学模型训练
2014 19th International Conference on Digital Signal Processing Pub Date : 2014-09-18 DOI: 10.1109/ICDSP.2014.6900717
Haofeng Kou, Weijia Shang
{"title":"Parallelized feature extraction and acoustic model training","authors":"Haofeng Kou, Weijia Shang","doi":"10.1109/ICDSP.2014.6900717","DOIUrl":"https://doi.org/10.1109/ICDSP.2014.6900717","url":null,"abstract":"In this paper, we present our research on the parallelized speech recognition including both Mel-Frequency Cepstral Coefficient (MFCC) feature extraction [1] and Viterbi training for Hidden Markov Model (HMM) based acoustic model [2] on the Graphics Processing Units (GPU). Robust and accurate speech recognition systems can only be realized with adequately trained acoustic models derived from the effectively parsed features. For common languages, state-of-the-art systems are extracted and trained on many thousands of hours of speech data and even with large clusters of machines the entire extracting and training process can take weeks. To overcome this development bottleneck, we not only demonstrate that feature extraction and acoustic model training are suitable for GPUs, but also propose the optimized parallel implementation using highly parallel GPUs by combining the MFCC feature extraction along with Viterbi training for HMM acoustic model, illustrate its application concurrency characteristics, data working set sizes, and describe the optimizations required for effective throughput on GPU processors. We demonstrate that feature extraction and acoustic model training are well suited for GPUs. Using one GTX580 our approach is shown to be overall approximately 95x faster than a sequential CPU implementation at the same accuracy level, enabling feature extraction and acoustic model training to be performed at realtime.","PeriodicalId":301856,"journal":{"name":"2014 19th International Conference on Digital Signal Processing","volume":"87 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127879335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Salient region detection for stereoscopic images 立体图像的显著区域检测
2014 19th International Conference on Digital Signal Processing Pub Date : 2014-09-18 DOI: 10.1109/ICDSP.2014.6900706
X. Fan, Zhi Liu, Guangling Sun
{"title":"Salient region detection for stereoscopic images","authors":"X. Fan, Zhi Liu, Guangling Sun","doi":"10.1109/ICDSP.2014.6900706","DOIUrl":"https://doi.org/10.1109/ICDSP.2014.6900706","url":null,"abstract":"In this paper, we propose an effective saliency model, which combines region-level depth, color and spatial information, to detect salient regions in stereoscopic images. Based on region segmentation results of stereoscopic images, depth contrast, depth weighted color contrast, and spatial compactness of color distribution are measured for each region, and combined to generate the region-level saliency map. Experimental results on a public stereoscopic image dataset with ground truths of salient objects demonstrate that the proposed saliency model outperforms the state-of-the-art saliency models.","PeriodicalId":301856,"journal":{"name":"2014 19th International Conference on Digital Signal Processing","volume":"6 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120923236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 71
A new score normalization for text-independent speaker verification 一种新的文本无关说话人验证的评分归一化
2014 19th International Conference on Digital Signal Processing Pub Date : 2014-09-18 DOI: 10.1109/ICDSP.2014.6900743
H. Ning, Y. Zou, Xuyan Hu
{"title":"A new score normalization for text-independent speaker verification","authors":"H. Ning, Y. Zou, Xuyan Hu","doi":"10.1109/ICDSP.2014.6900743","DOIUrl":"https://doi.org/10.1109/ICDSP.2014.6900743","url":null,"abstract":"In iVector-based speaker verification system, the claimed speaker was verified if the similarity between the iVector of the tested utterance (iVector-ts) and the iVector of the claimed speaker (iVector-cs) is smaller than a fixed threshold. The commonly used method to measure the similarity between the iVector-ts and iVector-cs is the cosine similarity scoring method. To further improve the performance of the speaker verification system when the training data is insufficient, a new scoring method termed as ratio normalization (Rnorm) scoring method is proposed, where the similarity between iVector-ts and iVector-cs is normalized by the dissimilarity between the tested speaker model and the universal background model (UBM). Preliminary experimental results with Timit database and self-built database show that our proposed Rnorm scoring method is able to reduce the equal error rate (EER) of the iVector-based TIV speaker verification system compared with that of using conventional cosine similarity scoring method.","PeriodicalId":301856,"journal":{"name":"2014 19th International Conference on Digital Signal Processing","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127118692","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Feature selection and channel optimization for biometric identification based on visual evoked potentials 基于视觉诱发电位的生物特征识别特征选择与通道优化
2014 19th International Conference on Digital Signal Processing Pub Date : 2014-09-18 DOI: 10.1109/ICDSP.2014.6900769
Yanru Bai, Zhiguo Zhang, Dong Ming
{"title":"Feature selection and channel optimization for biometric identification based on visual evoked potentials","authors":"Yanru Bai, Zhiguo Zhang, Dong Ming","doi":"10.1109/ICDSP.2014.6900769","DOIUrl":"https://doi.org/10.1109/ICDSP.2014.6900769","url":null,"abstract":"In recent years, biometric identification has received general concerns around the world, and become a frontal and hot topic in the information age. Among the internal biometric traits, electroencephalogram (EEG) signals have emerged as a prominent characteristic due to the high security, uniqueness and impossibility to steal or mimic. In this paper, individual difference of visual evoked potentials (VEPs) with cognition task were investigated, in addition, a feature selection and channel optimization strategy was developed for the VEPs based biometric identification system, where three different methods, including genetic algorithm (GA), Fisher discriminant ratio (FDR), and recursive feature elimination (RFE) were employed. In our experiments with 20 healthy subjects, the classification accuracy by support vector machine (SVM) reached up to 97.25% with AR model parameters, compared to 96.25% before optimization, and 32 channels of most discriminative were eventually selected from 64 channels with best performance. Results in this study revealed the feasibility of VEPs based EEG to be used for biometric identification. The proposed optimization algorithm was shown to have the ability to effectively improve the identification accuracy as well as simplifying the system. Further investigate may provide a novel idea for the individual difference analysis of EEG and for its practical design and optimization in the field of biometrics in the future.","PeriodicalId":301856,"journal":{"name":"2014 19th International Conference on Digital Signal Processing","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126579833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信