注视信息的多模态交互

Pernilla Qvarfordt
{"title":"注视信息的多模态交互","authors":"Pernilla Qvarfordt","doi":"10.1145/3015783.3015794","DOIUrl":null,"url":null,"abstract":"Observe a person pointing out and describing something. Where is that person looking? Chances are good that this person also looks at what she is talking about and pointing at. Gaze is naturally coordinated with our speech and hand movements. By utilizing this tendency, we can create a natural interaction with computing devices and environments. In multimodal gaze interaction, data from eye trackers are used as an active input mode, where, for instance, gaze is used as an alternative, or complementary, pointing modality along with other input modalities. Using gaze as an active, or explicit, input method is challenging for several reasons. One of them being that eyes are primarily used for perceiving our environment, so knowing when a person selects an item with gaze versus just looking around is an issue. Researchers have tried to solve this by combining gaze with various input methods, such as manual pointing, speech, touch, etc.However, gaze information can also be used in interactive systems for other purposes than explicit pointing, since a user's gaze is a good indication of the user's attention. In passive gaze interaction, the gaze is not used as the primary input method, but as a supporting one. In these kinds of systems, gaze is mainly used for inferring and reasoning about the user's cognitive state or activities in a way that can support the interaction. These kinds of multimodal systems often combine gaze with a multitude of input modalities. One example is to detect what features in an image a person is looking for, and use this information to suggest regions or other images that the person has not yet seen.In this chapter, eye movement and eye tracking data analysis is first reviewed (Section 9.2), followed by a discussion of eye movements in relation to other modalities (Section 9.3) to gain basic knowledge about eye tracking and gaze behavior. In Section 9.4, systems using gaze as active or passive input methods are discussed. Finally, Section 9.5 concludes the chapter.","PeriodicalId":222911,"journal":{"name":"The Handbook of Multimodal-Multisensor Interfaces, Volume 1","volume":"81 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"11","resultStr":"{\"title\":\"Gaze-informed multimodal interaction\",\"authors\":\"Pernilla Qvarfordt\",\"doi\":\"10.1145/3015783.3015794\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Observe a person pointing out and describing something. Where is that person looking? Chances are good that this person also looks at what she is talking about and pointing at. Gaze is naturally coordinated with our speech and hand movements. By utilizing this tendency, we can create a natural interaction with computing devices and environments. In multimodal gaze interaction, data from eye trackers are used as an active input mode, where, for instance, gaze is used as an alternative, or complementary, pointing modality along with other input modalities. Using gaze as an active, or explicit, input method is challenging for several reasons. One of them being that eyes are primarily used for perceiving our environment, so knowing when a person selects an item with gaze versus just looking around is an issue. Researchers have tried to solve this by combining gaze with various input methods, such as manual pointing, speech, touch, etc.However, gaze information can also be used in interactive systems for other purposes than explicit pointing, since a user's gaze is a good indication of the user's attention. In passive gaze interaction, the gaze is not used as the primary input method, but as a supporting one. In these kinds of systems, gaze is mainly used for inferring and reasoning about the user's cognitive state or activities in a way that can support the interaction. These kinds of multimodal systems often combine gaze with a multitude of input modalities. One example is to detect what features in an image a person is looking for, and use this information to suggest regions or other images that the person has not yet seen.In this chapter, eye movement and eye tracking data analysis is first reviewed (Section 9.2), followed by a discussion of eye movements in relation to other modalities (Section 9.3) to gain basic knowledge about eye tracking and gaze behavior. In Section 9.4, systems using gaze as active or passive input methods are discussed. Finally, Section 9.5 concludes the chapter.\",\"PeriodicalId\":222911,\"journal\":{\"name\":\"The Handbook of Multimodal-Multisensor Interfaces, Volume 1\",\"volume\":\"81 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2017-04-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"11\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"The Handbook of Multimodal-Multisensor Interfaces, Volume 1\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3015783.3015794\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"The Handbook of Multimodal-Multisensor Interfaces, Volume 1","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3015783.3015794","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 11

摘要

观察一个人指出并描述某事。那个人在看哪里?很有可能这个人也会看她在说什么,指着什么。目光自然地与我们的语言和手部动作相协调。通过利用这种趋势,我们可以创建与计算设备和环境的自然交互。在多模态凝视交互中,来自眼动仪的数据被用作主动输入模式,其中,例如,凝视与其他输入模式一起被用作替代或补充的指向模态。出于几个原因,将凝视作为一种主动或明确的输入法是具有挑战性的。其中之一是,眼睛主要用于感知我们的环境,所以知道一个人是在凝视中选择一件物品,还是只是环顾四周,这是一个问题。研究人员试图通过将凝视与各种输入方法相结合来解决这个问题,例如手动指向,语音,触摸等。然而,凝视信息也可以用于交互系统中的其他目的,而不是明确的指向,因为用户的凝视是用户注意力的良好指示。在被动注视交互中,注视不是作为主要输入法,而是作为辅助输入法。在这些类型的系统中,凝视主要用于以支持交互的方式推断和推理用户的认知状态或活动。这类多模态系统通常将凝视与多种输入模态结合起来。一个例子是检测一个人正在寻找图像中的哪些特征,并使用这些信息来建议这个人还没有看到的区域或其他图像。在本章中,首先回顾眼动和眼动追踪数据分析(第9.2节),然后讨论眼动与其他模态的关系(第9.3节),以获得有关眼动追踪和凝视行为的基本知识。在第9.4节中,将讨论使用注视作为主动或被动输入法的系统。最后,第9.5节结束本章。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Gaze-informed multimodal interaction
Observe a person pointing out and describing something. Where is that person looking? Chances are good that this person also looks at what she is talking about and pointing at. Gaze is naturally coordinated with our speech and hand movements. By utilizing this tendency, we can create a natural interaction with computing devices and environments. In multimodal gaze interaction, data from eye trackers are used as an active input mode, where, for instance, gaze is used as an alternative, or complementary, pointing modality along with other input modalities. Using gaze as an active, or explicit, input method is challenging for several reasons. One of them being that eyes are primarily used for perceiving our environment, so knowing when a person selects an item with gaze versus just looking around is an issue. Researchers have tried to solve this by combining gaze with various input methods, such as manual pointing, speech, touch, etc.However, gaze information can also be used in interactive systems for other purposes than explicit pointing, since a user's gaze is a good indication of the user's attention. In passive gaze interaction, the gaze is not used as the primary input method, but as a supporting one. In these kinds of systems, gaze is mainly used for inferring and reasoning about the user's cognitive state or activities in a way that can support the interaction. These kinds of multimodal systems often combine gaze with a multitude of input modalities. One example is to detect what features in an image a person is looking for, and use this information to suggest regions or other images that the person has not yet seen.In this chapter, eye movement and eye tracking data analysis is first reviewed (Section 9.2), followed by a discussion of eye movements in relation to other modalities (Section 9.3) to gain basic knowledge about eye tracking and gaze behavior. In Section 9.4, systems using gaze as active or passive input methods are discussed. Finally, Section 9.5 concludes the chapter.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信