基于感知特征的男女语音分类

Saptarshi Sengupta, Ghazaala Yasmin, Arijit Ghosal
{"title":"基于感知特征的男女语音分类","authors":"Saptarshi Sengupta, Ghazaala Yasmin, Arijit Ghosal","doi":"10.1109/ICCCNT.2017.8204065","DOIUrl":null,"url":null,"abstract":"Gender identification systems nowadays, are gaining momentum in terms of popularity because of their wide areas of application. They can be used in a variety of fields ranging from security and authentication services to content based information retrieval and also criminal investigations. Gender detection has started to gain importance because of the fact that recent studies conducted showed that the performance of gender dependent speech recognition models performs much better than gender independent models. In the proposed work, we aim to build such a system involving perceptual audio features such as pitch and tempo based features, short time energy etc., which are used to train classifiers to differentiate between the two classes of gender. We have selected such a combination of features as because previous works focused only on either pitch approach, MFCC approach etc., whereas our work is perhaps one of the first involving a combination of several such perceptual features. The system was tested on a wide range of speech files and was shown to be yielding promising results.","PeriodicalId":6581,"journal":{"name":"2017 8th International Conference on Computing, Communication and Networking Technologies (ICCCNT)","volume":"5 1","pages":"1-7"},"PeriodicalIF":0.0000,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":"{\"title\":\"Classification of male and female speech using perceptual features\",\"authors\":\"Saptarshi Sengupta, Ghazaala Yasmin, Arijit Ghosal\",\"doi\":\"10.1109/ICCCNT.2017.8204065\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Gender identification systems nowadays, are gaining momentum in terms of popularity because of their wide areas of application. They can be used in a variety of fields ranging from security and authentication services to content based information retrieval and also criminal investigations. Gender detection has started to gain importance because of the fact that recent studies conducted showed that the performance of gender dependent speech recognition models performs much better than gender independent models. In the proposed work, we aim to build such a system involving perceptual audio features such as pitch and tempo based features, short time energy etc., which are used to train classifiers to differentiate between the two classes of gender. We have selected such a combination of features as because previous works focused only on either pitch approach, MFCC approach etc., whereas our work is perhaps one of the first involving a combination of several such perceptual features. The system was tested on a wide range of speech files and was shown to be yielding promising results.\",\"PeriodicalId\":6581,\"journal\":{\"name\":\"2017 8th International Conference on Computing, Communication and Networking Technologies (ICCCNT)\",\"volume\":\"5 1\",\"pages\":\"1-7\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2017-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"6\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2017 8th International Conference on Computing, Communication and Networking Technologies (ICCCNT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCCNT.2017.8204065\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 8th International Conference on Computing, Communication and Networking Technologies (ICCCNT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCCNT.2017.8204065","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6

摘要

如今,性别识别系统由于其广泛的应用领域而越来越受欢迎。它们可以用于各种领域,从安全和身份验证服务到基于内容的信息检索以及刑事调查。性别检测已经开始变得越来越重要,因为最近的研究表明,性别依赖的语音识别模型的性能比性别独立的模型要好得多。在我们提出的工作中,我们的目标是建立这样一个系统,包括感知音频特征,如基于音高和节奏的特征,短时间能量等,这些特征用于训练分类器来区分两类性别。我们选择了这样的特征组合,因为以前的作品只关注音调方法,MFCC方法等,而我们的工作可能是第一个涉及几个这样的感知特征组合的工作之一。该系统在广泛的语音文件上进行了测试,并显示出令人满意的结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Classification of male and female speech using perceptual features
Gender identification systems nowadays, are gaining momentum in terms of popularity because of their wide areas of application. They can be used in a variety of fields ranging from security and authentication services to content based information retrieval and also criminal investigations. Gender detection has started to gain importance because of the fact that recent studies conducted showed that the performance of gender dependent speech recognition models performs much better than gender independent models. In the proposed work, we aim to build such a system involving perceptual audio features such as pitch and tempo based features, short time energy etc., which are used to train classifiers to differentiate between the two classes of gender. We have selected such a combination of features as because previous works focused only on either pitch approach, MFCC approach etc., whereas our work is perhaps one of the first involving a combination of several such perceptual features. The system was tested on a wide range of speech files and was shown to be yielding promising results.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信