{"title":"一种与NAO机器人进行方言语音交互的系统设计新方法","authors":"Ming Chen, Lujia Wang, Cheng-Zhong Xu, Renfa Li","doi":"10.1109/ICAR.2017.8023652","DOIUrl":null,"url":null,"abstract":"Intelligent human robot interaction are becoming popular in both industry and academia. However, amongst current techniques, speech recognition is a challenging topic, including real-time translation with high accuracy, amicability and the support for recognizing minor languages or sophisticated dialects. In this paper, we propose a human-friendly prototype deployed on NAO robots in a real-life scenario through daily speech commands and NAO would act accordingly. We primarily adopt HMM-GMM, the combination of HMMs (Hidden Markov Models) and GMMs (Gaussian Mixtures Models). The experimental results show that the proposed prototype achieves high accuracy and well-received by experiment subjects.","PeriodicalId":198633,"journal":{"name":"2017 18th International Conference on Advanced Robotics (ICAR)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":"{\"title\":\"A novel approach of system design for dialect speech interaction with NAO robot\",\"authors\":\"Ming Chen, Lujia Wang, Cheng-Zhong Xu, Renfa Li\",\"doi\":\"10.1109/ICAR.2017.8023652\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Intelligent human robot interaction are becoming popular in both industry and academia. However, amongst current techniques, speech recognition is a challenging topic, including real-time translation with high accuracy, amicability and the support for recognizing minor languages or sophisticated dialects. In this paper, we propose a human-friendly prototype deployed on NAO robots in a real-life scenario through daily speech commands and NAO would act accordingly. We primarily adopt HMM-GMM, the combination of HMMs (Hidden Markov Models) and GMMs (Gaussian Mixtures Models). The experimental results show that the proposed prototype achieves high accuracy and well-received by experiment subjects.\",\"PeriodicalId\":198633,\"journal\":{\"name\":\"2017 18th International Conference on Advanced Robotics (ICAR)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2017-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"6\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2017 18th International Conference on Advanced Robotics (ICAR)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICAR.2017.8023652\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 18th International Conference on Advanced Robotics (ICAR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAR.2017.8023652","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A novel approach of system design for dialect speech interaction with NAO robot
Intelligent human robot interaction are becoming popular in both industry and academia. However, amongst current techniques, speech recognition is a challenging topic, including real-time translation with high accuracy, amicability and the support for recognizing minor languages or sophisticated dialects. In this paper, we propose a human-friendly prototype deployed on NAO robots in a real-life scenario through daily speech commands and NAO would act accordingly. We primarily adopt HMM-GMM, the combination of HMMs (Hidden Markov Models) and GMMs (Gaussian Mixtures Models). The experimental results show that the proposed prototype achieves high accuracy and well-received by experiment subjects.