Webcam-based categorization of task engagement of PC users at work

T. Ohara, Nobuyuki Umezu
{"title":"Webcam-based categorization of task engagement of PC users at work","authors":"T. Ohara, Nobuyuki Umezu","doi":"10.1117/12.2589104","DOIUrl":null,"url":null,"abstract":"In this paper, we propose a support method for PC users to monitor their task engagement. Our approach is based on the number of keyboard strokes and pixel changes on the screen, and images from an ordinary webcam equipped with a PC. With our system, supervisors or teachers would improve their quality of guidance and instructions given at the right moment when workers or students require some support due to reasons such as slow progress and technical difficulties. In conventional methods, a special device such as an acceleration sensor for each individual is often required to acquire information on one’s working status and body movements, which is difficult to deploy in a real environment due to its cost for sensors. A face detection method based on Deep Neural Network, such as SSD, allows as to implement a cheaper system using an ordinary web camera. We calculate the average difference between two grabbed frames from the user’s screen to estimate the amount of screen changes between a given time interval. The number of key strokes typed by the user is another factor to estimate their task engagement. These factors are used to categorize the work mode of users. We use the K-Means method based on the Euclidean distance to cluster the recorded factors to determine thresholds for task categorization. We conducted experiments seven participants to evaluate the accuracy of our categorization method. Every participant is asked to categorize 15 scenes into four work modes. A scene includes a camera image with the PC user’s face, the screenshot path the moment, and the number of key strokes. The results from these participants are then compared with those of our system that categorized the same scenes with the thresholds on three factors. Approximately only 60% of these results matched each other, where we have enough room to improve our approach. Future work includes the selection of features that are far more effective for categorization, a better estimation of pixel changes on the PC screen, and evaluation experiments with more participants.","PeriodicalId":295011,"journal":{"name":"International Conference on Quality Control by Artificial Vision","volume":"7 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Conference on Quality Control by Artificial Vision","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.2589104","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In this paper, we propose a support method for PC users to monitor their task engagement. Our approach is based on the number of keyboard strokes and pixel changes on the screen, and images from an ordinary webcam equipped with a PC. With our system, supervisors or teachers would improve their quality of guidance and instructions given at the right moment when workers or students require some support due to reasons such as slow progress and technical difficulties. In conventional methods, a special device such as an acceleration sensor for each individual is often required to acquire information on one’s working status and body movements, which is difficult to deploy in a real environment due to its cost for sensors. A face detection method based on Deep Neural Network, such as SSD, allows as to implement a cheaper system using an ordinary web camera. We calculate the average difference between two grabbed frames from the user’s screen to estimate the amount of screen changes between a given time interval. The number of key strokes typed by the user is another factor to estimate their task engagement. These factors are used to categorize the work mode of users. We use the K-Means method based on the Euclidean distance to cluster the recorded factors to determine thresholds for task categorization. We conducted experiments seven participants to evaluate the accuracy of our categorization method. Every participant is asked to categorize 15 scenes into four work modes. A scene includes a camera image with the PC user’s face, the screenshot path the moment, and the number of key strokes. The results from these participants are then compared with those of our system that categorized the same scenes with the thresholds on three factors. Approximately only 60% of these results matched each other, where we have enough room to improve our approach. Future work includes the selection of features that are far more effective for categorization, a better estimation of pixel changes on the PC screen, and evaluation experiments with more participants.
基于网络摄像头的PC用户工作任务参与分类
在本文中,我们提出了一种支持PC用户监控他们的任务参与的方法。我们的方法是基于键盘敲击的次数和屏幕上的像素变化,以及配备PC的普通网络摄像头的图像。有了我们的系统,当工人或学生由于进度缓慢或技术困难等原因需要一些支持时,主管或教师可以在适当的时候提高指导和指导的质量。在传统的方法中,通常需要为每个人配备一个特殊的设备,如加速度传感器,以获取个人的工作状态和身体运动信息,由于传感器成本高,难以在真实环境中部署。一种基于深度神经网络(Deep Neural Network)的人脸检测方法,如SSD,可以使用普通的网络摄像头实现更便宜的系统。我们计算从用户屏幕上抓取的两帧之间的平均差值,以估计给定时间间隔内屏幕变化的量。用户敲击键盘的次数是评估他们对任务投入程度的另一个因素。这些因素被用来对用户的工作模式进行分类。我们使用基于欧氏距离的K-Means方法对记录的因素进行聚类,以确定任务分类的阈值。我们进行了7个参与者的实验来评估我们的分类方法的准确性。每位参与者被要求将15个场景分为四种工作模式。场景包括带有PC用户脸部的相机图像、瞬间的截图路径和按键次数。然后将这些参与者的结果与我们的系统的结果进行比较,我们的系统根据三个因素的阈值对相同的场景进行分类。大约只有60%的结果相互匹配,我们有足够的空间来改进我们的方法。未来的工作包括选择更有效的分类特征,更好地估计PC屏幕上的像素变化,以及有更多参与者的评估实验。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信