{"title":"SRA System Design: Using Deep Learning to Analyse Experimental data for Speech Researchers","authors":"Tianshi Xie, Dallin J Bailey, Cheryl D. Seals","doi":"10.1109/ICTACSE50438.2022.10009746","DOIUrl":null,"url":null,"abstract":"Speech researchers expend tremendous effort when measuring and analyzing audio data from experimental participants and evaluating large amounts of audio data to aid speech research. In traditional assessment methods, researchers listen verbatim to long audio files from participants to find the keywords needed for the assessment. This approach is very tedious and time intensive for speech researchers. To improve research efficiency and efficacy, we designed a system called SRA (Speech Recognition Assistant) based on the DeepSpeech model. SRA effectively supports speech researchers in the evaluation of participant experimental audio data. The purpose is to achieve the following: (i) simplify the workflow for analyzing data and (ii) reduce the time cost for researchers to extract data; (iii) provide a framework with potential for adaptation for use in other fields, such as speech science, audiology, and hearing science.","PeriodicalId":301767,"journal":{"name":"2022 International Conference on Theoretical and Applied Computer Science and Engineering (ICTASCE)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Theoretical and Applied Computer Science and Engineering (ICTASCE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICTACSE50438.2022.10009746","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Speech researchers expend tremendous effort when measuring and analyzing audio data from experimental participants and evaluating large amounts of audio data to aid speech research. In traditional assessment methods, researchers listen verbatim to long audio files from participants to find the keywords needed for the assessment. This approach is very tedious and time intensive for speech researchers. To improve research efficiency and efficacy, we designed a system called SRA (Speech Recognition Assistant) based on the DeepSpeech model. SRA effectively supports speech researchers in the evaluation of participant experimental audio data. The purpose is to achieve the following: (i) simplify the workflow for analyzing data and (ii) reduce the time cost for researchers to extract data; (iii) provide a framework with potential for adaptation for use in other fields, such as speech science, audiology, and hearing science.