Multimodal Dialogue Management for Multiparty Interaction with Infants

Setareh Nasihati Gilani, D. Traum, A. Merla, Eugenia Hee, Zoey Walker, Barbara Manini, Grady Gallagher, L. Petitto
{"title":"Multimodal Dialogue Management for Multiparty Interaction with Infants","authors":"Setareh Nasihati Gilani, D. Traum, A. Merla, Eugenia Hee, Zoey Walker, Barbara Manini, Grady Gallagher, L. Petitto","doi":"10.1145/3242969.3243029","DOIUrl":null,"url":null,"abstract":"We present dialogue management routines for a system to engage in multiparty agent-infant interaction. The ultimate purpose of this research is to help infants learn a visual sign language by engaging them in naturalistic and socially contingent conversations during an early-life critical period for language development (ages 6 to 12 months) as initiated by an artificial agent. As a first step, we focus on creating and maintaining agent-infant engagement that elicits appropriate and socially contingent responses from the baby. Our system includes two agents, a physical robot and an animated virtual human. The system's multimodal perception includes an eye-tracker (measures attention) and a thermal infrared imaging camera (measures patterns of emotional arousal). A dialogue policy is presented that selects individual actions and planned multiparty sequences based on perceptual inputs about the baby's internal changing states of emotional engagement. The present version of the system was evaluated in interaction with 8 babies. All babies demonstrated spontaneous and sustained engagement with the agents for several minutes, with patterns of conversationally relevant and socially contingent behaviors. We further performed a detailed case-study analysis with annotation of all agent and baby behaviors. Results show that the baby's behaviors were generally relevant to agent conversations and contained direct evidence for socially contingent responses by the baby to specific linguistic samples produced by the avatar. This work demonstrates the potential for language learning from agents in very young babies and has especially broad implications regarding the use of artificial agents with babies who have minimal language exposure in early life.","PeriodicalId":308751,"journal":{"name":"Proceedings of the 20th ACM International Conference on Multimodal Interaction","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"21","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 20th ACM International Conference on Multimodal Interaction","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3242969.3243029","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 21

Abstract

We present dialogue management routines for a system to engage in multiparty agent-infant interaction. The ultimate purpose of this research is to help infants learn a visual sign language by engaging them in naturalistic and socially contingent conversations during an early-life critical period for language development (ages 6 to 12 months) as initiated by an artificial agent. As a first step, we focus on creating and maintaining agent-infant engagement that elicits appropriate and socially contingent responses from the baby. Our system includes two agents, a physical robot and an animated virtual human. The system's multimodal perception includes an eye-tracker (measures attention) and a thermal infrared imaging camera (measures patterns of emotional arousal). A dialogue policy is presented that selects individual actions and planned multiparty sequences based on perceptual inputs about the baby's internal changing states of emotional engagement. The present version of the system was evaluated in interaction with 8 babies. All babies demonstrated spontaneous and sustained engagement with the agents for several minutes, with patterns of conversationally relevant and socially contingent behaviors. We further performed a detailed case-study analysis with annotation of all agent and baby behaviors. Results show that the baby's behaviors were generally relevant to agent conversations and contained direct evidence for socially contingent responses by the baby to specific linguistic samples produced by the avatar. This work demonstrates the potential for language learning from agents in very young babies and has especially broad implications regarding the use of artificial agents with babies who have minimal language exposure in early life.
婴儿多方互动的多模式对话管理
我们提出了一个系统的对话管理例程,以参与多方代理-婴儿交互。本研究的最终目的是帮助婴儿在生命早期语言发展的关键时期(6至12个月)通过人工代理发起的自然和社会偶然对话来学习视觉手语。作为第一步,我们专注于创造和维持代理-婴儿的参与,从婴儿引出适当的和社会偶然的反应。我们的系统包括两个代理,一个实体机器人和一个动画虚拟人。该系统的多模态感知包括一个眼球追踪器(测量注意力)和一个热红外成像相机(测量情绪唤醒模式)。提出了一种对话策略,该策略根据婴儿情绪参与的内部变化状态的感知输入来选择个人行为和计划的多方序列。目前版本的系统在与8个婴儿的互动中进行了评估。所有的婴儿都表现出自发的和持续几分钟的与代理人的接触,具有对话相关和社会偶然行为的模式。我们进一步进行了详细的案例研究分析,并注释了所有代理和婴儿的行为。结果表明,婴儿的行为通常与代理对话相关,并包含婴儿对虚拟化身产生的特定语言样本的社会偶然反应的直接证据。这项工作证明了在非常年幼的婴儿中从代理学习语言的潜力,并且对于在早期生活中语言接触最少的婴儿使用人工代理具有特别广泛的意义。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信