测试、测量和自动语音识别

ACM Stand. Pub Date : 1997-09-01 DOI:10.1145/266231.266238
D. S. Pallett, J. Baker
{"title":"测试、测量和自动语音识别","authors":"D. S. Pallett, J. Baker","doi":"10.1145/266231.266238","DOIUrl":null,"url":null,"abstract":"igure One shows a representative test cycle for tests implemented by the NIST group. A test cycle is initiated with an analysis and planning phase, typically coordinated by a group of researchers, research sponsors, and NIST staff. During this phase, test protocols and implementation schedules are defined. A data-collection phase leads to the creation or identification of standardized speech and natural language corpora, distributed to a community of core technology developers. In most cases, a portion of the corpora is held in reserve by NIST as performance assessment test material. At agreed-upon times, NIST defines and releases development and evaluation test sets to the core technology developers, and they, in turn, provide NIST with the results of their locally-implemented tests. NIST then produces a detailed set of uniformly-scored tabulated results, including the results of numerous paired-comparison statistical significance tests and other analyses. These test results and their scientific implications then become an important matter for discussion at technical meetings. The extent of NIST's work is illustrated by a look at some 60 technical papers on speech recognition submitted to the 1996 IEEE International Conference on Acoustics, Speech and Signal Processing. Twenty-eight of the 60 papers reported results based on the use of NIST-defined test data, test methodologies, and NIST-implemented benchmark tests. Of these 28 papers, 16 were by researchers in the United States and 12 were from other nations. From Dragon Systems' perspective, the NIST reference speech database measurement and testing methodologies are important for research and necessary to advance the technology. While ideas are plentiful , testing is expensive; researchers and research resources are costly. So sharing data makes sense. Large common databases are statistically more meaningful than smaller proprietary ones, and using these large databases minimizes dead-end approaches. At the speech-recognition workshops where results of the NIST's benchmark tests are presented, there are opportunities to compare results and the different approaches pursued at different laboratories. In this way the entire community benefits.","PeriodicalId":270594,"journal":{"name":"ACM Stand.","volume":"78 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1997-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Tests, measurements, and automatic speech recognition\",\"authors\":\"D. S. Pallett, J. Baker\",\"doi\":\"10.1145/266231.266238\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"igure One shows a representative test cycle for tests implemented by the NIST group. A test cycle is initiated with an analysis and planning phase, typically coordinated by a group of researchers, research sponsors, and NIST staff. During this phase, test protocols and implementation schedules are defined. A data-collection phase leads to the creation or identification of standardized speech and natural language corpora, distributed to a community of core technology developers. In most cases, a portion of the corpora is held in reserve by NIST as performance assessment test material. At agreed-upon times, NIST defines and releases development and evaluation test sets to the core technology developers, and they, in turn, provide NIST with the results of their locally-implemented tests. NIST then produces a detailed set of uniformly-scored tabulated results, including the results of numerous paired-comparison statistical significance tests and other analyses. These test results and their scientific implications then become an important matter for discussion at technical meetings. The extent of NIST's work is illustrated by a look at some 60 technical papers on speech recognition submitted to the 1996 IEEE International Conference on Acoustics, Speech and Signal Processing. Twenty-eight of the 60 papers reported results based on the use of NIST-defined test data, test methodologies, and NIST-implemented benchmark tests. Of these 28 papers, 16 were by researchers in the United States and 12 were from other nations. From Dragon Systems' perspective, the NIST reference speech database measurement and testing methodologies are important for research and necessary to advance the technology. While ideas are plentiful , testing is expensive; researchers and research resources are costly. So sharing data makes sense. Large common databases are statistically more meaningful than smaller proprietary ones, and using these large databases minimizes dead-end approaches. At the speech-recognition workshops where results of the NIST's benchmark tests are presented, there are opportunities to compare results and the different approaches pursued at different laboratories. In this way the entire community benefits.\",\"PeriodicalId\":270594,\"journal\":{\"name\":\"ACM Stand.\",\"volume\":\"78 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1997-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ACM Stand.\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/266231.266238\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACM Stand.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/266231.266238","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

摘要

图1显示了NIST组实现的测试的代表性测试周期。测试周期从分析和计划阶段开始,通常由一组研究人员、研究赞助者和NIST工作人员进行协调。在此阶段,将定义测试协议和实现时间表。数据收集阶段导致创建或识别标准化的语音和自然语言语料库,并将其分发给核心技术开发人员社区。在大多数情况下,一部分语料库由NIST保留作为性能评估测试材料。在商定的时间,NIST定义并向核心技术开发人员发布开发和评估测试集,而他们反过来向NIST提供他们在本地实现的测试结果。然后,NIST生成一组详细的统一评分的表格结果,包括许多配对比较统计显著性测试和其他分析的结果。这些测试结果及其科学含义随后成为技术会议讨论的重要问题。从1996年IEEE声学、语音和信号处理国际会议上提交的60多篇关于语音识别的技术论文中可以看出NIST的工作范围。60篇论文中有28篇报告了基于使用nist定义的测试数据、测试方法和nist实现的基准测试的结果。在这28篇论文中,有16篇来自美国,12篇来自其他国家。从Dragon Systems的角度来看,NIST参考语音数据库测量和测试方法对于研究非常重要,并且对于推进该技术是必要的。虽然想法很多,但测试是昂贵的;研究人员和研究资源是昂贵的。所以共享数据是有意义的。大型公共数据库在统计上比小型专有数据库更有意义,并且使用这些大型数据库可以最大限度地减少死胡同。在展示NIST基准测试结果的语音识别研讨会上,有机会比较结果和不同实验室采用的不同方法。这样整个社区都会受益。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Tests, measurements, and automatic speech recognition
igure One shows a representative test cycle for tests implemented by the NIST group. A test cycle is initiated with an analysis and planning phase, typically coordinated by a group of researchers, research sponsors, and NIST staff. During this phase, test protocols and implementation schedules are defined. A data-collection phase leads to the creation or identification of standardized speech and natural language corpora, distributed to a community of core technology developers. In most cases, a portion of the corpora is held in reserve by NIST as performance assessment test material. At agreed-upon times, NIST defines and releases development and evaluation test sets to the core technology developers, and they, in turn, provide NIST with the results of their locally-implemented tests. NIST then produces a detailed set of uniformly-scored tabulated results, including the results of numerous paired-comparison statistical significance tests and other analyses. These test results and their scientific implications then become an important matter for discussion at technical meetings. The extent of NIST's work is illustrated by a look at some 60 technical papers on speech recognition submitted to the 1996 IEEE International Conference on Acoustics, Speech and Signal Processing. Twenty-eight of the 60 papers reported results based on the use of NIST-defined test data, test methodologies, and NIST-implemented benchmark tests. Of these 28 papers, 16 were by researchers in the United States and 12 were from other nations. From Dragon Systems' perspective, the NIST reference speech database measurement and testing methodologies are important for research and necessary to advance the technology. While ideas are plentiful , testing is expensive; researchers and research resources are costly. So sharing data makes sense. Large common databases are statistically more meaningful than smaller proprietary ones, and using these large databases minimizes dead-end approaches. At the speech-recognition workshops where results of the NIST's benchmark tests are presented, there are opportunities to compare results and the different approaches pursued at different laboratories. In this way the entire community benefits.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信