Swapnil Bhosale, I. Sheikh, Sri Harsha Dumpala, S. Kopparapu
{"title":"低资源口语理解的迁移学习","authors":"Swapnil Bhosale, I. Sheikh, Sri Harsha Dumpala, S. Kopparapu","doi":"10.1109/IBSSC47189.2019.8973067","DOIUrl":null,"url":null,"abstract":"Spoken Language Understanding (SLU) without speech-to-text conversion is more promising in low resource scenarios. These could be applications where there is not enough labeled data to train reliable speech recognition and language understanding systems, or where running SLU on edge is preferred over cloud based services. In this paper, we present an approach for building SLU without speech-to-text conversion in low resource scenarios using a transfer learning approach. We show that the intermediate layer representations from a pre-trained model outperforms the typically used Mel filter bank features. Moreover, the representations extracted from a model pre-trained on one language perform well even for SLU tasks on a different language.","PeriodicalId":148941,"journal":{"name":"2019 IEEE Bombay Section Signature Conference (IBSSC)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Transfer Learning for Low Resource Spoken Language Understanding without Speech-to-Text\",\"authors\":\"Swapnil Bhosale, I. Sheikh, Sri Harsha Dumpala, S. Kopparapu\",\"doi\":\"10.1109/IBSSC47189.2019.8973067\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Spoken Language Understanding (SLU) without speech-to-text conversion is more promising in low resource scenarios. These could be applications where there is not enough labeled data to train reliable speech recognition and language understanding systems, or where running SLU on edge is preferred over cloud based services. In this paper, we present an approach for building SLU without speech-to-text conversion in low resource scenarios using a transfer learning approach. We show that the intermediate layer representations from a pre-trained model outperforms the typically used Mel filter bank features. Moreover, the representations extracted from a model pre-trained on one language perform well even for SLU tasks on a different language.\",\"PeriodicalId\":148941,\"journal\":{\"name\":\"2019 IEEE Bombay Section Signature Conference (IBSSC)\",\"volume\":\"2 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 IEEE Bombay Section Signature Conference (IBSSC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IBSSC47189.2019.8973067\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE Bombay Section Signature Conference (IBSSC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IBSSC47189.2019.8973067","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Transfer Learning for Low Resource Spoken Language Understanding without Speech-to-Text
Spoken Language Understanding (SLU) without speech-to-text conversion is more promising in low resource scenarios. These could be applications where there is not enough labeled data to train reliable speech recognition and language understanding systems, or where running SLU on edge is preferred over cloud based services. In this paper, we present an approach for building SLU without speech-to-text conversion in low resource scenarios using a transfer learning approach. We show that the intermediate layer representations from a pre-trained model outperforms the typically used Mel filter bank features. Moreover, the representations extracted from a model pre-trained on one language perform well even for SLU tasks on a different language.