Saurabhchand Bhati, Chunxi Liu, J. Villalba, J. Trmal, S. Khudanpur, N. Dehak
{"title":"基于声学单元的自下而上无监督词发现","authors":"Saurabhchand Bhati, Chunxi Liu, J. Villalba, J. Trmal, S. Khudanpur, N. Dehak","doi":"10.1109/GlobalSIP45357.2019.8969225","DOIUrl":null,"url":null,"abstract":"Unsupervised term discovery is the task of identifying and grouping reoccurring word-like patterns from the untranscribed audio data. It facilitates unsupervised acoustic model training in zero resource setting where no or minimal transcribed speech is available. In this paper, we investigate two-step bottom-up approaches for unsupervised discovery of word-like units. The first step discovers phone-like acoustic units from data and the second step combines the basic acoustic blocks to identify word-like units. We investigated Embedded Segmental K-means and Nested Hierarchical Pitman-Yor (PYR) model as bottom-up strategies. ESK-Means iteratively selects boundaries from an initial set to arrive at the word boundaries. The final performance critically depends on the quality of the initial boundaries. We used a segmentation method that discovers boundaries much closer to actual boundaries. PYR model has been used for word segmentation from space removed text data, and here we use it for word discovery from unsupervised acoustic units. The term discovery performance is evaluated on the Zero Resource 2017 challenge dataset, which consists of around 70 hours of unlabelled data. Our systems outperformed the baseline systems on all the languages without language-specific parameter tuning. We performed comprehensive experiments of the system parameters on the system performance.","PeriodicalId":221378,"journal":{"name":"2019 IEEE Global Conference on Signal and Information Processing (GlobalSIP)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Bottom-Up Unsupervised Word Discovery via Acoustic Units\",\"authors\":\"Saurabhchand Bhati, Chunxi Liu, J. Villalba, J. Trmal, S. Khudanpur, N. Dehak\",\"doi\":\"10.1109/GlobalSIP45357.2019.8969225\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Unsupervised term discovery is the task of identifying and grouping reoccurring word-like patterns from the untranscribed audio data. It facilitates unsupervised acoustic model training in zero resource setting where no or minimal transcribed speech is available. In this paper, we investigate two-step bottom-up approaches for unsupervised discovery of word-like units. The first step discovers phone-like acoustic units from data and the second step combines the basic acoustic blocks to identify word-like units. We investigated Embedded Segmental K-means and Nested Hierarchical Pitman-Yor (PYR) model as bottom-up strategies. ESK-Means iteratively selects boundaries from an initial set to arrive at the word boundaries. The final performance critically depends on the quality of the initial boundaries. We used a segmentation method that discovers boundaries much closer to actual boundaries. PYR model has been used for word segmentation from space removed text data, and here we use it for word discovery from unsupervised acoustic units. The term discovery performance is evaluated on the Zero Resource 2017 challenge dataset, which consists of around 70 hours of unlabelled data. Our systems outperformed the baseline systems on all the languages without language-specific parameter tuning. We performed comprehensive experiments of the system parameters on the system performance.\",\"PeriodicalId\":221378,\"journal\":{\"name\":\"2019 IEEE Global Conference on Signal and Information Processing (GlobalSIP)\",\"volume\":\"8 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 IEEE Global Conference on Signal and Information Processing (GlobalSIP)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/GlobalSIP45357.2019.8969225\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE Global Conference on Signal and Information Processing (GlobalSIP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/GlobalSIP45357.2019.8969225","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Bottom-Up Unsupervised Word Discovery via Acoustic Units
Unsupervised term discovery is the task of identifying and grouping reoccurring word-like patterns from the untranscribed audio data. It facilitates unsupervised acoustic model training in zero resource setting where no or minimal transcribed speech is available. In this paper, we investigate two-step bottom-up approaches for unsupervised discovery of word-like units. The first step discovers phone-like acoustic units from data and the second step combines the basic acoustic blocks to identify word-like units. We investigated Embedded Segmental K-means and Nested Hierarchical Pitman-Yor (PYR) model as bottom-up strategies. ESK-Means iteratively selects boundaries from an initial set to arrive at the word boundaries. The final performance critically depends on the quality of the initial boundaries. We used a segmentation method that discovers boundaries much closer to actual boundaries. PYR model has been used for word segmentation from space removed text data, and here we use it for word discovery from unsupervised acoustic units. The term discovery performance is evaluated on the Zero Resource 2017 challenge dataset, which consists of around 70 hours of unlabelled data. Our systems outperformed the baseline systems on all the languages without language-specific parameter tuning. We performed comprehensive experiments of the system parameters on the system performance.