Masashi Inoue, M. Ogihara, Ryoko Hanada, N. Furuyama
{"title":"Utility of Gestural Cues in Indexing Semantic Miscommunication","authors":"Masashi Inoue, M. Ogihara, Ryoko Hanada, N. Furuyama","doi":"10.1109/FUTURETECH.2010.5482653","DOIUrl":null,"url":null,"abstract":"In multimedia data analysis, automated indexing of conversational video is an emerging topic. One challenging problem in this topic is the recognition of higher-level concepts, such as \\emph{miscommunications} in conversations. While detecting these miscommunications is generally easy for the speakers as well as for observers, it is not currently understood which cues contribute to their detection and to what extent. We investigate the possibility of indexing the occurrence of miscommunications in psychotherapeutic face-to-face conversations from gestural patterns. The applicability of machine learning is investigated as a means of detecting miscommunication from gestural patterns observed in the conversations. Both simple and complex classifiers are constructed using different features taken from the gesture data. Both short-term and long-term effects are tested using different time window sizes. Also, two types of gestures, communicative and non-communicative, are considered. The experimental results suggest that there \\emph{does not exist a single gestural feature} that can explain the occurrence of semantic miscommunication. Another interesting finding is that gestural cues correlate more with long-term gestural patterns than short-term ones.","PeriodicalId":380192,"journal":{"name":"2010 5th International Conference on Future Information Technology","volume":"4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 5th International Conference on Future Information Technology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/FUTURETECH.2010.5482653","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
In multimedia data analysis, automated indexing of conversational video is an emerging topic. One challenging problem in this topic is the recognition of higher-level concepts, such as \emph{miscommunications} in conversations. While detecting these miscommunications is generally easy for the speakers as well as for observers, it is not currently understood which cues contribute to their detection and to what extent. We investigate the possibility of indexing the occurrence of miscommunications in psychotherapeutic face-to-face conversations from gestural patterns. The applicability of machine learning is investigated as a means of detecting miscommunication from gestural patterns observed in the conversations. Both simple and complex classifiers are constructed using different features taken from the gesture data. Both short-term and long-term effects are tested using different time window sizes. Also, two types of gestures, communicative and non-communicative, are considered. The experimental results suggest that there \emph{does not exist a single gestural feature} that can explain the occurrence of semantic miscommunication. Another interesting finding is that gestural cues correlate more with long-term gestural patterns than short-term ones.