Vision-based engagement detection in Virtual Reality

Ghassem Tofighi, Haisong Gu, K. Raahemifar
{"title":"Vision-based engagement detection in Virtual Reality","authors":"Ghassem Tofighi, Haisong Gu, K. Raahemifar","doi":"10.1109/DMIAF.2016.7574933","DOIUrl":null,"url":null,"abstract":"User engagement modeling for manipulating actions in vision-based interfaces is one of the most important case studies of user mental state detection. In a Virtual Reality environment that employs camera sensors to recognize human activities, we have to know were user intend to perform an action and when he/she is disengaged. Without a proper algorithm for recognizing engagement status, any kind of activities could be interpreted as manipulating actions, called “Midas Touch” problem. Baseline approach for solving this problem is activating gesture recognition system using some focus gestures such as waiving or raising hand. However, a desirable natural user interface should be able to understand user's mental status automatically. In this paper, a novel multi-modal model for engagement detection, DAIA 1, is presented. using DAIA, the spectrum of mental status for performing an action is quantized in a finite number of engagement states. For this purpose, a Finite State Transducer (FST) is designed. This engagement framework shows how to integrate multi-modal information from user biometric data streams such as 2D and 3D imaging. FST is employed to make the state transition smoothly using combination of several boolean expressions. Our FST true detection rate is 92.3% in total for four different states. Results also show FST can segment user hand gestures more robustly.","PeriodicalId":404025,"journal":{"name":"2016 Digital Media Industry & Academic Forum (DMIAF)","volume":"67 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 Digital Media Industry & Academic Forum (DMIAF)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/DMIAF.2016.7574933","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 8

Abstract

User engagement modeling for manipulating actions in vision-based interfaces is one of the most important case studies of user mental state detection. In a Virtual Reality environment that employs camera sensors to recognize human activities, we have to know were user intend to perform an action and when he/she is disengaged. Without a proper algorithm for recognizing engagement status, any kind of activities could be interpreted as manipulating actions, called “Midas Touch” problem. Baseline approach for solving this problem is activating gesture recognition system using some focus gestures such as waiving or raising hand. However, a desirable natural user interface should be able to understand user's mental status automatically. In this paper, a novel multi-modal model for engagement detection, DAIA 1, is presented. using DAIA, the spectrum of mental status for performing an action is quantized in a finite number of engagement states. For this purpose, a Finite State Transducer (FST) is designed. This engagement framework shows how to integrate multi-modal information from user biometric data streams such as 2D and 3D imaging. FST is employed to make the state transition smoothly using combination of several boolean expressions. Our FST true detection rate is 92.3% in total for four different states. Results also show FST can segment user hand gestures more robustly.
虚拟现实中基于视觉的交战检测
在基于视觉的界面中,用户参与建模是用户心理状态检测最重要的案例之一。在使用摄像头传感器来识别人类活动的虚拟现实环境中,我们必须知道用户是否打算执行操作,以及他/她何时脱离操作。如果没有适当的算法来识别参与状态,任何类型的活动都可能被解释为操纵行为,称为“点石成金”问题。解决这个问题的基本方法是使用一些焦点手势(如放弃或举手)来激活手势识别系统。然而,一个理想的自然用户界面应该能够自动理解用户的心理状态。本文提出了一种新的多模态碰撞检测模型daia1。使用DAIA,执行一个动作的心理状态谱被量化为有限数量的参与状态。为此,设计了有限状态传感器(FST)。该参与框架展示了如何整合来自用户生物特征数据流(如2D和3D成像)的多模态信息。利用FST将多个布尔表达式组合在一起,实现状态平滑转换。我们对四种不同状态的FST真实检出率为92.3%。结果还表明,FST可以更稳健地分割用户手势。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信