An Adaptive Q Cochlear Filter in Phoneme Recognition

T. Hirahara
{"title":"An Adaptive Q Cochlear Filter in Phoneme Recognition","authors":"T. Hirahara","doi":"10.1109/ASPAA.1991.634091","DOIUrl":null,"url":null,"abstract":"Introduction It has been expected that speech recognition performance can be improved by replacing a traditional front-end with a model of the auditory periphery. The underlying assumption is that if a model could be designed properly, it should generate a more efficient representation compared to traditional physical spectrum representations. From this viewpoint, several works have been reported [11-[41. However, these studies do not always show an auditory front-end to be superior to a traditional front-end. Some auditory front-ends are superior only for noisy speech, but many show little, if any, superiority in processing clean speech. We also have been developing an auditory model characterized by an adaptive Q cochlear filter not only for the front-end of a speech recognition system but also for a general purpose spectral1 analyzer in speech research [SI. In this paper, several auditory front-ends based on the adaptive Q cochlear filter and its relatives are tested in speaker dependent phoneme recognition using different stochastic pattern classifier!;, a shift invariant multi template matching system using LVQ2-trained codebook, and a VQ-HMM system. Further, we will discuss problems of using an auditory model as a frontend of an automatic speech recognition system. 2. Adaptive Q Cochlear Filter An adaptive Q cochlear filter (AQF) is a computational filter that functionally simulates the nonlinear filtering characteristics of the basilar membrane vibrating system. The AQF consists of three parts: (1) cascaded second-order notch filters (NOTCH), (2) second-order band pass filters (BPF) connected to each NOTCH output, and (3) adaptive Q circuits connected to each BPF output. The adaptive Q circuit consists of a second-order low-pass filter (LPF) in","PeriodicalId":146017,"journal":{"name":"Final Program and Paper Summaries 1991 IEEE ASSP Workshop on Applications of Signal Processing to Audio and Acoustics","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Final Program and Paper Summaries 1991 IEEE ASSP Workshop on Applications of Signal Processing to Audio and Acoustics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ASPAA.1991.634091","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Introduction It has been expected that speech recognition performance can be improved by replacing a traditional front-end with a model of the auditory periphery. The underlying assumption is that if a model could be designed properly, it should generate a more efficient representation compared to traditional physical spectrum representations. From this viewpoint, several works have been reported [11-[41. However, these studies do not always show an auditory front-end to be superior to a traditional front-end. Some auditory front-ends are superior only for noisy speech, but many show little, if any, superiority in processing clean speech. We also have been developing an auditory model characterized by an adaptive Q cochlear filter not only for the front-end of a speech recognition system but also for a general purpose spectral1 analyzer in speech research [SI. In this paper, several auditory front-ends based on the adaptive Q cochlear filter and its relatives are tested in speaker dependent phoneme recognition using different stochastic pattern classifier!;, a shift invariant multi template matching system using LVQ2-trained codebook, and a VQ-HMM system. Further, we will discuss problems of using an auditory model as a frontend of an automatic speech recognition system. 2. Adaptive Q Cochlear Filter An adaptive Q cochlear filter (AQF) is a computational filter that functionally simulates the nonlinear filtering characteristics of the basilar membrane vibrating system. The AQF consists of three parts: (1) cascaded second-order notch filters (NOTCH), (2) second-order band pass filters (BPF) connected to each NOTCH output, and (3) adaptive Q circuits connected to each BPF output. The adaptive Q circuit consists of a second-order low-pass filter (LPF) in
一种用于音素识别的自适应Q耳蜗滤波器
人们一直期望通过用听觉外围模型取代传统的前端来提高语音识别性能。潜在的假设是,如果一个模型可以被适当地设计,它应该比传统的物理频谱表示产生更有效的表示。从这一观点出发,已经报道了一些作品[11-[41]。然而,这些研究并不总是表明听觉前端优于传统前端。一些听觉前端只对嘈杂的语音有优势,但许多在处理干净的语音方面几乎没有优势。我们还开发了一种以自适应Q耳蜗滤波器为特征的听觉模型,不仅用于语音识别系统的前端,也用于语音研究中的通用频谱分析仪[SI]。本文采用不同的随机模式分类器,对基于自适应Q耳蜗滤波器及其相关听觉前端进行了基于说话人的音素识别测试;采用lvq2训练码本的移位不变多模板匹配系统,以及VQ-HMM系统。此外,我们将讨论使用听觉模型作为自动语音识别系统前端的问题。2. 自适应Q耳蜗滤波器(Adaptive Q Filter, AQF)是一种模拟基底膜振动系统非线性滤波特性的计算滤波器。AQF由三部分组成:(1)级联二阶陷波滤波器(notch),(2)连接到每个notch输出端的二阶带通滤波器(BPF),以及(3)连接到每个BPF输出端的自适应Q电路。自适应Q电路由一个二阶低通滤波器(LPF)组成
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信