Head movement during natural group conversation and inter-annotator agreement on manual annotation

Angkana Lertpoompunya, Nathan C Higgins, Erol J. Ozmeral, D. Eddins
{"title":"Head movement during natural group conversation and inter-annotator agreement on manual annotation","authors":"Angkana Lertpoompunya, Nathan C Higgins, Erol J. Ozmeral, D. Eddins","doi":"10.1121/10.0022958","DOIUrl":null,"url":null,"abstract":"During speech communication and conversational turn-taking, listeners direct their head and eyes to receive meaningful auditory and visual cues. Features of these behaviors may convey listener intent. This study designed a test environment, data collection protocol and procedures, and investigated head movement behaviors during self-driven conversations among multiple partners. Nine participants were tested in cohorts of three. Participants wore a headset with sensors tracked by an infrared camera system. Participants watched an audio-video clip, followed by a 5-min undirected discussion. The entire session was video recorded for annotation purposes. Two annotators independently coded the video files using the EUDICO Linguistic Annotator software application. Annotations were then co-registered with the head tracking data in post processing. Inter-annotator agreement demonstrated the desired reliability, thereby validating the procedures designed. Movement trajectories showed that there were individual differences on the head yaw distribution. The combination of objective measures of head movement and manual annotation of conversation behaviors provides a rich data set for characterizing natural conversations in ecologically valid settings. The measurement procedures and coding system developed here is a first step towards characterizing head movements during conversations needed to predict listening intent and to create actions based on those predictions.","PeriodicalId":256727,"journal":{"name":"The Journal of the Acoustical Society of America","volume":"30 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"The Journal of the Acoustical Society of America","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1121/10.0022958","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

During speech communication and conversational turn-taking, listeners direct their head and eyes to receive meaningful auditory and visual cues. Features of these behaviors may convey listener intent. This study designed a test environment, data collection protocol and procedures, and investigated head movement behaviors during self-driven conversations among multiple partners. Nine participants were tested in cohorts of three. Participants wore a headset with sensors tracked by an infrared camera system. Participants watched an audio-video clip, followed by a 5-min undirected discussion. The entire session was video recorded for annotation purposes. Two annotators independently coded the video files using the EUDICO Linguistic Annotator software application. Annotations were then co-registered with the head tracking data in post processing. Inter-annotator agreement demonstrated the desired reliability, thereby validating the procedures designed. Movement trajectories showed that there were individual differences on the head yaw distribution. The combination of objective measures of head movement and manual annotation of conversation behaviors provides a rich data set for characterizing natural conversations in ecologically valid settings. The measurement procedures and coding system developed here is a first step towards characterizing head movements during conversations needed to predict listening intent and to create actions based on those predictions.
小组自然交谈时的头部运动和人工标注时标注者之间的一致意见
在语言交流和会话轮番进行的过程中,听者会引导头部和眼睛接收有意义的听觉和视觉提示。这些行为的特征可能传达了听者的意图。本研究设计了一个测试环境、数据收集协议和程序,并调查了多人对话中自我驱动的头部运动行为。九名参与者每三人一组接受了测试。参与者佩戴的耳机带有红外摄像系统跟踪的传感器。参与者先观看一段音频视频片段,然后进行 5 分钟的无指导讨论。整个过程都进行了录像,以便进行注释。两名注释员使用 EUDICO 语言注释器软件应用程序对视频文件进行独立编码。然后在后期处理中将注释与头部跟踪数据共同注册。注释者之间的一致性达到了预期的可靠性,从而验证了所设计的程序。运动轨迹显示,头部偏航分布存在个体差异。头部运动的客观测量和对话行为的人工标注相结合,为在生态有效的环境中描述自然对话提供了丰富的数据集。本文所开发的测量程序和编码系统是表征对话过程中头部运动特征的第一步,可用于预测倾听意图,并根据这些预测制定行动。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信