S. Srinivasan, D. Ponceleón, D. Petkovic, M. Viswanathan
{"title":"Query expansion for imperfect speech: applications in distributed learning","authors":"S. Srinivasan, D. Ponceleón, D. Petkovic, M. Viswanathan","doi":"10.1109/IVL.2000.853839","DOIUrl":null,"url":null,"abstract":"Advances in speech recognition technology have shown encouraging results for spoken document retrieval where the average precision often approaches 70% of that achieved for perfect text transcriptions. Typical applications of spoken document retrieval pertain to retrieval of stories from archived video/audio assets. In the CueVideo project, our application focus is spoken document retrieval from a video database for just-in-time training/distributed learning. Typical content is not pre-segmented, has no predefined structure, is of varying audio quality, and may not have domain specific data available. For such content, we propose a two level search, namely, a first level search across the entire video collection, and a second level search within a specific video. At both search levels, we perform an experimental evaluation of a combination of new and existing query expansion methods, intended to offset retrieval errors due to misrecognition.","PeriodicalId":333664,"journal":{"name":"2000 Proceedings Workshop on Content-based Access of Image and Video Libraries","volume":"14 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2000-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"12","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2000 Proceedings Workshop on Content-based Access of Image and Video Libraries","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IVL.2000.853839","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 12
Abstract
Advances in speech recognition technology have shown encouraging results for spoken document retrieval where the average precision often approaches 70% of that achieved for perfect text transcriptions. Typical applications of spoken document retrieval pertain to retrieval of stories from archived video/audio assets. In the CueVideo project, our application focus is spoken document retrieval from a video database for just-in-time training/distributed learning. Typical content is not pre-segmented, has no predefined structure, is of varying audio quality, and may not have domain specific data available. For such content, we propose a two level search, namely, a first level search across the entire video collection, and a second level search within a specific video. At both search levels, we perform an experimental evaluation of a combination of new and existing query expansion methods, intended to offset retrieval errors due to misrecognition.