{"title":"机器学习连接语音视听记录的发音数据","authors":"R. Trencsényi, L. Czap","doi":"10.1109/CITDS54976.2022.9914284","DOIUrl":null,"url":null,"abstract":"The center of attraction of the present study is the application of neural networks for combining data arising from dynamic audiovisual sources made by ultrasound and magnetic resonance imaging methods, which store image and sound signals recorded during human speech. The objectives of machine learning are tongue contours fitted to the frames of the audiovisual packages by automatic contour tracking algorithms.","PeriodicalId":271992,"journal":{"name":"2022 IEEE 2nd Conference on Information Technology and Data Science (CITDS)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Articulatory Data of Audiovisual Records of Speech Connected by Machine Learning\",\"authors\":\"R. Trencsényi, L. Czap\",\"doi\":\"10.1109/CITDS54976.2022.9914284\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The center of attraction of the present study is the application of neural networks for combining data arising from dynamic audiovisual sources made by ultrasound and magnetic resonance imaging methods, which store image and sound signals recorded during human speech. The objectives of machine learning are tongue contours fitted to the frames of the audiovisual packages by automatic contour tracking algorithms.\",\"PeriodicalId\":271992,\"journal\":{\"name\":\"2022 IEEE 2nd Conference on Information Technology and Data Science (CITDS)\",\"volume\":\"35 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-05-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE 2nd Conference on Information Technology and Data Science (CITDS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CITDS54976.2022.9914284\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 2nd Conference on Information Technology and Data Science (CITDS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CITDS54976.2022.9914284","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Articulatory Data of Audiovisual Records of Speech Connected by Machine Learning
The center of attraction of the present study is the application of neural networks for combining data arising from dynamic audiovisual sources made by ultrasound and magnetic resonance imaging methods, which store image and sound signals recorded during human speech. The objectives of machine learning are tongue contours fitted to the frames of the audiovisual packages by automatic contour tracking algorithms.