发音训练中的发音运动传感

Aslan B. Wong, Xia Chen, Qianru Liao, Kaishun Wu
{"title":"发音训练中的发音运动传感","authors":"Aslan B. Wong, Xia Chen, Qianru Liao, Kaishun Wu","doi":"10.1109/SECON52354.2021.9491610","DOIUrl":null,"url":null,"abstract":"The vowel is deemed the essence of the syllable in which controls the articulation of each word uttered. However, articulation sensing has not been adequately evaluated. The challenging task is that the speech signal contains insufficient information for articulation analysis. We propose a new approach to identify the articulation of monophthongs in multiple languages. We employ simultaneously two ranges of acoustic signals, both speech and ultrasonic signal, to recognize lip shape and tongue position, which is implemented into an off-the-shelf smartphone to be more accessible. The articulation recognition accuracy is 94.74%. The proposed system also applies to an alternative model for a pronunciation training system that gives articulation feedback to a user.","PeriodicalId":120945,"journal":{"name":"2021 18th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON)","volume":"68 1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Articulation Motion Sensing for Pronunciation Training\",\"authors\":\"Aslan B. Wong, Xia Chen, Qianru Liao, Kaishun Wu\",\"doi\":\"10.1109/SECON52354.2021.9491610\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The vowel is deemed the essence of the syllable in which controls the articulation of each word uttered. However, articulation sensing has not been adequately evaluated. The challenging task is that the speech signal contains insufficient information for articulation analysis. We propose a new approach to identify the articulation of monophthongs in multiple languages. We employ simultaneously two ranges of acoustic signals, both speech and ultrasonic signal, to recognize lip shape and tongue position, which is implemented into an off-the-shelf smartphone to be more accessible. The articulation recognition accuracy is 94.74%. The proposed system also applies to an alternative model for a pronunciation training system that gives articulation feedback to a user.\",\"PeriodicalId\":120945,\"journal\":{\"name\":\"2021 18th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON)\",\"volume\":\"68 1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-07-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 18th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SECON52354.2021.9491610\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 18th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SECON52354.2021.9491610","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

元音被认为是音节的本质,它控制着每个单词的发音。然而,关节感知尚未得到充分的评估。具有挑战性的任务是语音信号包含的信息不足,不足以进行发音分析。我们提出了一种新的方法来识别多语言中单音的发音。我们同时使用两种声音信号,语音和超声波信号,来识别嘴唇的形状和舌头的位置,这是实现到一个现成的智能手机更容易访问。发音识别准确率为94.74%。所提出的系统也适用于发音训练系统的替代模型,该模型向用户提供发音反馈。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Articulation Motion Sensing for Pronunciation Training
The vowel is deemed the essence of the syllable in which controls the articulation of each word uttered. However, articulation sensing has not been adequately evaluated. The challenging task is that the speech signal contains insufficient information for articulation analysis. We propose a new approach to identify the articulation of monophthongs in multiple languages. We employ simultaneously two ranges of acoustic signals, both speech and ultrasonic signal, to recognize lip shape and tongue position, which is implemented into an off-the-shelf smartphone to be more accessible. The articulation recognition accuracy is 94.74%. The proposed system also applies to an alternative model for a pronunciation training system that gives articulation feedback to a user.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信