{"title":"发音训练中的发音运动传感","authors":"Aslan B. Wong, Xia Chen, Qianru Liao, Kaishun Wu","doi":"10.1109/SECON52354.2021.9491610","DOIUrl":null,"url":null,"abstract":"The vowel is deemed the essence of the syllable in which controls the articulation of each word uttered. However, articulation sensing has not been adequately evaluated. The challenging task is that the speech signal contains insufficient information for articulation analysis. We propose a new approach to identify the articulation of monophthongs in multiple languages. We employ simultaneously two ranges of acoustic signals, both speech and ultrasonic signal, to recognize lip shape and tongue position, which is implemented into an off-the-shelf smartphone to be more accessible. The articulation recognition accuracy is 94.74%. The proposed system also applies to an alternative model for a pronunciation training system that gives articulation feedback to a user.","PeriodicalId":120945,"journal":{"name":"2021 18th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON)","volume":"68 1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Articulation Motion Sensing for Pronunciation Training\",\"authors\":\"Aslan B. Wong, Xia Chen, Qianru Liao, Kaishun Wu\",\"doi\":\"10.1109/SECON52354.2021.9491610\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The vowel is deemed the essence of the syllable in which controls the articulation of each word uttered. However, articulation sensing has not been adequately evaluated. The challenging task is that the speech signal contains insufficient information for articulation analysis. We propose a new approach to identify the articulation of monophthongs in multiple languages. We employ simultaneously two ranges of acoustic signals, both speech and ultrasonic signal, to recognize lip shape and tongue position, which is implemented into an off-the-shelf smartphone to be more accessible. The articulation recognition accuracy is 94.74%. The proposed system also applies to an alternative model for a pronunciation training system that gives articulation feedback to a user.\",\"PeriodicalId\":120945,\"journal\":{\"name\":\"2021 18th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON)\",\"volume\":\"68 1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-07-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 18th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SECON52354.2021.9491610\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 18th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SECON52354.2021.9491610","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Articulation Motion Sensing for Pronunciation Training
The vowel is deemed the essence of the syllable in which controls the articulation of each word uttered. However, articulation sensing has not been adequately evaluated. The challenging task is that the speech signal contains insufficient information for articulation analysis. We propose a new approach to identify the articulation of monophthongs in multiple languages. We employ simultaneously two ranges of acoustic signals, both speech and ultrasonic signal, to recognize lip shape and tongue position, which is implemented into an off-the-shelf smartphone to be more accessible. The articulation recognition accuracy is 94.74%. The proposed system also applies to an alternative model for a pronunciation training system that gives articulation feedback to a user.