Appearance-Based Gaze Tracking Through Supervised Machine Learning

Daniel Melesse, Mahmoud Khalil, Elias Kagabo, T. Ning, Kevin Huang
{"title":"Appearance-Based Gaze Tracking Through Supervised Machine Learning","authors":"Daniel Melesse, Mahmoud Khalil, Elias Kagabo, T. Ning, Kevin Huang","doi":"10.1109/ICSP48669.2020.9321075","DOIUrl":null,"url":null,"abstract":"Applications that use human gaze have become increasingly more popular in the domain of human-computer interfaces, and advances in eye gaze tracking technology over the past few decades have led to the development of promising gaze estimation techniques. In this paper, a low-cost, in-house video camera-based gaze tracking system was developed, trained and evaluated. Seminal gaze detection methods constrained the application space to indoor conditions, and in most cases techniques required intrusive hardware. More modern gaze detection techniques try to eliminate the use of any additional hardware to reduce monetary cost as well as undue burden to the user, all the while maintaining accuracy of detection. In this work, image acquisition was achieved using a low-cost USB web camera mounted at a fixed position on the viewing screen or laptop. In order to determine the point of gaze, the Viola Jones face detection algorithm is used to extract facial features from the image frame. The gaze is then calculated using image processing techniques to extract gaze features, namely related to the image position of the pupil. Thousands of images are classified and labeled to form an in-house database. A multi-class Support Vector Machine (SVM) was trained and tested on this data set to distinguish point of gaze from input face image. Cross validation was used to train the model. Confusion matrices, accuracy, precision, and recall are used to evaluate the performance of the classification model. Evaluation of the proposed appearance-based technique using two different kernel functions is also assessed in detail.","PeriodicalId":237073,"journal":{"name":"2020 15th IEEE International Conference on Signal Processing (ICSP)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 15th IEEE International Conference on Signal Processing (ICSP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSP48669.2020.9321075","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Applications that use human gaze have become increasingly more popular in the domain of human-computer interfaces, and advances in eye gaze tracking technology over the past few decades have led to the development of promising gaze estimation techniques. In this paper, a low-cost, in-house video camera-based gaze tracking system was developed, trained and evaluated. Seminal gaze detection methods constrained the application space to indoor conditions, and in most cases techniques required intrusive hardware. More modern gaze detection techniques try to eliminate the use of any additional hardware to reduce monetary cost as well as undue burden to the user, all the while maintaining accuracy of detection. In this work, image acquisition was achieved using a low-cost USB web camera mounted at a fixed position on the viewing screen or laptop. In order to determine the point of gaze, the Viola Jones face detection algorithm is used to extract facial features from the image frame. The gaze is then calculated using image processing techniques to extract gaze features, namely related to the image position of the pupil. Thousands of images are classified and labeled to form an in-house database. A multi-class Support Vector Machine (SVM) was trained and tested on this data set to distinguish point of gaze from input face image. Cross validation was used to train the model. Confusion matrices, accuracy, precision, and recall are used to evaluate the performance of the classification model. Evaluation of the proposed appearance-based technique using two different kernel functions is also assessed in detail.
通过监督机器学习的基于外观的凝视跟踪
在人机界面领域,使用人眼注视的应用越来越受欢迎,在过去的几十年里,眼球注视跟踪技术的进步导致了有前途的注视估计技术的发展。本文开发了一种低成本的、基于内部摄像机的注视跟踪系统,并对其进行了训练和评估。开创性的凝视检测方法将应用空间限制在室内条件下,并且在大多数情况下技术需要侵入性硬件。更现代的注视检测技术试图消除任何额外硬件的使用,以减少金钱成本以及对用户的不必要负担,同时保持检测的准确性。在这项工作中,图像采集是通过安装在观看屏幕或笔记本电脑固定位置的低成本USB网络摄像头实现的。为了确定凝视点,采用Viola Jones人脸检测算法从图像帧中提取人脸特征。然后使用图像处理技术计算凝视,提取凝视特征,即与瞳孔的图像位置相关。成千上万的图像被分类和标记,形成一个内部数据库。在此数据集上训练并测试了一种多类支持向量机(SVM),用于从输入的人脸图像中识别注视点。采用交叉验证对模型进行训练。混淆矩阵、准确度、精密度和召回率被用来评估分类模型的性能。使用两种不同的核函数对所提出的基于外观的技术进行了详细的评估。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信