{"title":"Articulatory Data of Audiovisual Records of Speech Connected by Machine Learning","authors":"R. Trencsényi, L. Czap","doi":"10.1109/CITDS54976.2022.9914284","DOIUrl":null,"url":null,"abstract":"The center of attraction of the present study is the application of neural networks for combining data arising from dynamic audiovisual sources made by ultrasound and magnetic resonance imaging methods, which store image and sound signals recorded during human speech. The objectives of machine learning are tongue contours fitted to the frames of the audiovisual packages by automatic contour tracking algorithms.","PeriodicalId":271992,"journal":{"name":"2022 IEEE 2nd Conference on Information Technology and Data Science (CITDS)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 2nd Conference on Information Technology and Data Science (CITDS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CITDS54976.2022.9914284","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The center of attraction of the present study is the application of neural networks for combining data arising from dynamic audiovisual sources made by ultrasound and magnetic resonance imaging methods, which store image and sound signals recorded during human speech. The objectives of machine learning are tongue contours fitted to the frames of the audiovisual packages by automatic contour tracking algorithms.