{"title":"Interactive retrieval of spoken content optimizing by LDA algorithm","authors":"S. Chavan, R. Kagalkar","doi":"10.1109/I2C2.2017.8321791","DOIUrl":null,"url":null,"abstract":"Communication between the human are performed by using different languages. Language is either in written or spoken form. In our paper, as an author we are representing a framework that gives output as a description for audio file using signal processing. The output is in the text format which is derived by examining the audio content and providing the description of audio in the text format. So the study of translation of audio into text goes increasing. The framework is distributed into two sections called training and testing section. The training section is train the audio with its description like speech conversation present in that audio. This data is stored in the system using database with features of scenario of audio. Another section is testing section. The testing section test the audio file and retrieve the output as description of audio comparing audios stored into database (i.e. in training section). Using Latent Dirichlet allocation processing sentences are generated from objects and their activities.","PeriodicalId":288351,"journal":{"name":"2017 International Conference on Intelligent Computing and Control (I2C2)","volume":"54 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 International Conference on Intelligent Computing and Control (I2C2)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/I2C2.2017.8321791","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Communication between the human are performed by using different languages. Language is either in written or spoken form. In our paper, as an author we are representing a framework that gives output as a description for audio file using signal processing. The output is in the text format which is derived by examining the audio content and providing the description of audio in the text format. So the study of translation of audio into text goes increasing. The framework is distributed into two sections called training and testing section. The training section is train the audio with its description like speech conversation present in that audio. This data is stored in the system using database with features of scenario of audio. Another section is testing section. The testing section test the audio file and retrieve the output as description of audio comparing audios stored into database (i.e. in training section). Using Latent Dirichlet allocation processing sentences are generated from objects and their activities.