Christopher D. McMurrough, Jonathan Rich, C. Conly, V. Athitsos, F. Makedon
{"title":"Multi-modal object of interest detection using eye gaze and RGB-D cameras","authors":"Christopher D. McMurrough, Jonathan Rich, C. Conly, V. Athitsos, F. Makedon","doi":"10.1145/2401836.2401838","DOIUrl":null,"url":null,"abstract":"This paper presents a low-cost, wearable headset for mobile 3D Point of Gaze (PoG) estimation in assistive applications. The device consists of an eye tracking camera and forward facing RGB-D scene camera which are able to provide an estimate of the user gaze vector and its intersection with a 3D point in space. A computational approach that considers object 3D information and visual appearance together with the visual gaze interactions of the user is also given to demonstrate the utility of the device. The resulting system is able to identify, in real-time, known objects within a scene that intersect with the user gaze vector.","PeriodicalId":272657,"journal":{"name":"Gaze-In '12","volume":"543 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2012-10-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Gaze-In '12","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/2401836.2401838","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
This paper presents a low-cost, wearable headset for mobile 3D Point of Gaze (PoG) estimation in assistive applications. The device consists of an eye tracking camera and forward facing RGB-D scene camera which are able to provide an estimate of the user gaze vector and its intersection with a 3D point in space. A computational approach that considers object 3D information and visual appearance together with the visual gaze interactions of the user is also given to demonstrate the utility of the device. The resulting system is able to identify, in real-time, known objects within a scene that intersect with the user gaze vector.