{"title":"Effectiveness of Feature Collaboration in Speaker Identification for Voice Biometrics","authors":"Arunima Das, L. P. Roy, Santos Kumar Das","doi":"10.1109/ICCECE51049.2023.10085318","DOIUrl":null,"url":null,"abstract":"Voice biometrics is a promising solution to online banking that doesn’t need one’s physical presence, unlike fingerprint and retina scanners. Systems for identifying speakers are a crucial component of biometric technologies. Over the past few years, numerous speaker identification systems have been developed and used; these systems rely on various feature extraction methodologies. Due to its capacity to capture’ the repeated nature and effectiveness of signals, short-time characteristics like perceptual linear predictive (PLP) and Mel frequency cepstral coefficients (MFCC) have been used in the majority of studies on speaker identification. The efficiency of MFCC characteristics in accurately identifying speakers has been demonstrated in various research. However, the’ performance of these features degrades in noisy environments. To address this feature, a novel feature fusion of some spectral and time-domain features has been suggested in this paper. Moreover, this study evaluates the effectiveness of feature collaboration for speaker identification. The experimental results show that the suggested feature vector and classifying model can be widely applied to different types of voice biometric systems.","PeriodicalId":447131,"journal":{"name":"2023 International Conference on Computer, Electrical & Communication Engineering (ICCECE)","volume":"80 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 International Conference on Computer, Electrical & Communication Engineering (ICCECE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCECE51049.2023.10085318","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Voice biometrics is a promising solution to online banking that doesn’t need one’s physical presence, unlike fingerprint and retina scanners. Systems for identifying speakers are a crucial component of biometric technologies. Over the past few years, numerous speaker identification systems have been developed and used; these systems rely on various feature extraction methodologies. Due to its capacity to capture’ the repeated nature and effectiveness of signals, short-time characteristics like perceptual linear predictive (PLP) and Mel frequency cepstral coefficients (MFCC) have been used in the majority of studies on speaker identification. The efficiency of MFCC characteristics in accurately identifying speakers has been demonstrated in various research. However, the’ performance of these features degrades in noisy environments. To address this feature, a novel feature fusion of some spectral and time-domain features has been suggested in this paper. Moreover, this study evaluates the effectiveness of feature collaboration for speaker identification. The experimental results show that the suggested feature vector and classifying model can be widely applied to different types of voice biometric systems.