Communicating Eye-gaze Across a Distance: Comparing an Eye-gaze enabled Immersive Collaborative Virtual Environment, Aligned Video Conferencing, and Being Together

D. Roberts, R. Wolff, John P Rae, A. Steed, R. Aspin, Moira McIntyre, Adriana Pena, Oyewole Oyekoya, W. Steptoe
{"title":"Communicating Eye-gaze Across a Distance: Comparing an Eye-gaze enabled Immersive Collaborative Virtual Environment, Aligned Video Conferencing, and Being Together","authors":"D. Roberts, R. Wolff, John P Rae, A. Steed, R. Aspin, Moira McIntyre, Adriana Pena, Oyewole Oyekoya, W. Steptoe","doi":"10.1109/VR.2009.4811013","DOIUrl":null,"url":null,"abstract":"Eye gaze is an important and widely studied non-verbal resource in co-located social interaction. When we attempt to support tele-presence between people, there are two main technologies that can be used today: video-conferencing (VC) and collaborative virtual environments (CVEs). In VC, one can observe eye-gaze behaviour but practically the targets of eye-gaze are only correct if the participants remain relatively still. We attempt to support eye-gaze behaviour in an unconstrained manner by integrating eye-trackers into an Immersive CVE (ICVE) system. This paper aims to show that while both ICVE and VC allow people to discern being looked at and what else is looked at, when someone gazes into their space from another location, ICVE alone can continue to do this as people move. The conditions of aligned VC, ICVE, eye-gaze enabled ICVE and co-location are compared. The impact of factors of alignment, lighting, resolution, and perspective distortion are minimised through a set of pilot experiments, before a formal experiment records results for optimal settings. Results show that both VC and ICVE support eye-gaze in constrained situations, but only ICVE supports movement of the observer. We quantify the mis-judgements that are made and discuss how our findings might inform research into supporting eye-gaze through interpolated free viewpoint video based methods.","PeriodicalId":433266,"journal":{"name":"2009 IEEE Virtual Reality Conference","volume":"66 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2009-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"47","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2009 IEEE Virtual Reality Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/VR.2009.4811013","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 47

Abstract

Eye gaze is an important and widely studied non-verbal resource in co-located social interaction. When we attempt to support tele-presence between people, there are two main technologies that can be used today: video-conferencing (VC) and collaborative virtual environments (CVEs). In VC, one can observe eye-gaze behaviour but practically the targets of eye-gaze are only correct if the participants remain relatively still. We attempt to support eye-gaze behaviour in an unconstrained manner by integrating eye-trackers into an Immersive CVE (ICVE) system. This paper aims to show that while both ICVE and VC allow people to discern being looked at and what else is looked at, when someone gazes into their space from another location, ICVE alone can continue to do this as people move. The conditions of aligned VC, ICVE, eye-gaze enabled ICVE and co-location are compared. The impact of factors of alignment, lighting, resolution, and perspective distortion are minimised through a set of pilot experiments, before a formal experiment records results for optimal settings. Results show that both VC and ICVE support eye-gaze in constrained situations, but only ICVE supports movement of the observer. We quantify the mis-judgements that are made and discuss how our findings might inform research into supporting eye-gaze through interpolated free viewpoint video based methods.
远距离交流眼神:比较眼神支持的沉浸式协作虚拟环境、对齐视频会议和在一起
眼神注视是一种重要的非语言资源,在同地社会交往中被广泛研究。当我们试图支持人与人之间的远程呈现时,有两种主要的技术可以使用:视频会议(VC)和协作虚拟环境(CVEs)。在VC中,人们可以观察到眼睛注视的行为,但实际上眼睛注视的目标只有在参与者保持相对静止的情况下才是正确的。我们试图通过将眼动追踪器集成到沉浸式CVE (ICVE)系统中,以不受约束的方式支持眼球注视行为。本文旨在表明,虽然ICVE和VC都能让人们辨别被看和被看的东西,但当有人从另一个位置凝视他们的空间时,ICVE可以在人们移动时继续这样做。比较了对齐VC、ICVE、眼注视启用ICVE和共定位的条件。在正式的实验记录最佳设置结果之前,通过一组试点实验,将对齐、照明、分辨率和透视失真等因素的影响降至最低。结果表明,在受限情况下,VC和ICVE都支持人眼注视,但只有ICVE支持观察者的运动。我们量化了所做的错误判断,并讨论了我们的发现如何通过基于插值的免费视点视频方法为支持眼睛凝视的研究提供信息。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信