Calibration-free and deep-learning-based customer gaze direction detection technology based on the YOLOv3-tiny model for smart advertising displays

IF 1 4区 工程技术 Q3 ENGINEERING, MULTIDISCIPLINARY
Wei-Liang Ou, Yu-Hsiu Cheng, Chin-Chieh Chang, Hua-Luen Chen, Chih-Peng Fan
{"title":"Calibration-free and deep-learning-based customer gaze direction detection technology based on the YOLOv3-tiny model for smart advertising displays","authors":"Wei-Liang Ou, Yu-Hsiu Cheng, Chin-Chieh Chang, Hua-Luen Chen, Chih-Peng Fan","doi":"10.1080/02533839.2023.2262724","DOIUrl":null,"url":null,"abstract":"ABSTRACTBecause of the COVID-19 pandemic, gaze tracking for nontouch user interface designs used in advertising displays or automatic vending machines has become an emerging research topic. In this study, a cost-effective deep-learning-based customer gaze direction detection technology was developed for a smart advertising display. To achieve calibration-free interactions between customers and displays, the You-Only-Look-Once (YOLO)-v3-tiny-based deep learning model was used for determining the bounding boxes of eyes and pupils. Next, postprocessing was conducted using a voting mechanism and difference vectors between the central coordinates of the bounding boxes for effectively predicting customer gaze directions. Product images were separated into two or four gaze zones. For cross-person testing, the Recall, Precision, Accuracy, and F1-score for two gaze zones were approximately 77%, 99%, 88%, and 87%, respectively, and those for four gaze zones were approximately 72%, 91%, 91%, and 79%, respectively. Software implementations on NVIDIA graphics-processing-unit-accelerated embedded platforms exhibited a frame rate of nearly 30 frames per second. The proposed design achieved real-time gaze direction detection for a smart advertising platform.CO EDITOR-IN-CHIEF: Yuan, Shyan-MingASSOCIATE EDITOR: Yuan, Shyan-MingKEYWORDS: Deep learningYOLOv3-tinyintelligent systemssmart displaysnontouch user interface designgaze direction detectioncalibration-free Nomenclature UL=the gaze state estimated at the upper left directionUR=the gaze state estimated at the upper right directionDL=the gaze state estimated at the down left directionDR=the gaze state estimated1 at the down right directionC_pupil=the central coordinate position of the right or left pupilC_eye=the central coordinate position of the right or left eyeV_d=the difference vector between two central coordinate positionsX1=the central coordinate position of X-axis of the pupil’s bounding boxY1=the central coordinate position of Y-axis of the pupil’s bounding boxX2=the central coordinate position of X-axis of the eye’s bounding boxY2=the central coordinate position of Y-axis of the eye’s bounding boxTN=the number of true negative casesTP=the number of true positive casesFN=the number of false negative casesFP=the number of false positive casesF1 Score=it is a measure of a test’s accuracy by using 2×Precision×Recall/(Precision + Recall)mAP=it is a metric used to measure the performance of models doing object detection tasksDisclosure statementNo potential conflict of interest was reported by the author(s).Additional informationFundingThis work was financially supported by the Ministry of Science and Technology (MOST) under Grant No. [109-2218-E-005-008].","PeriodicalId":17313,"journal":{"name":"Journal of the Chinese Institute of Engineers","volume":"75 1","pages":"0"},"PeriodicalIF":1.0000,"publicationDate":"2023-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of the Chinese Institute of Engineers","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1080/02533839.2023.2262724","RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

Abstract

ABSTRACTBecause of the COVID-19 pandemic, gaze tracking for nontouch user interface designs used in advertising displays or automatic vending machines has become an emerging research topic. In this study, a cost-effective deep-learning-based customer gaze direction detection technology was developed for a smart advertising display. To achieve calibration-free interactions between customers and displays, the You-Only-Look-Once (YOLO)-v3-tiny-based deep learning model was used for determining the bounding boxes of eyes and pupils. Next, postprocessing was conducted using a voting mechanism and difference vectors between the central coordinates of the bounding boxes for effectively predicting customer gaze directions. Product images were separated into two or four gaze zones. For cross-person testing, the Recall, Precision, Accuracy, and F1-score for two gaze zones were approximately 77%, 99%, 88%, and 87%, respectively, and those for four gaze zones were approximately 72%, 91%, 91%, and 79%, respectively. Software implementations on NVIDIA graphics-processing-unit-accelerated embedded platforms exhibited a frame rate of nearly 30 frames per second. The proposed design achieved real-time gaze direction detection for a smart advertising platform.CO EDITOR-IN-CHIEF: Yuan, Shyan-MingASSOCIATE EDITOR: Yuan, Shyan-MingKEYWORDS: Deep learningYOLOv3-tinyintelligent systemssmart displaysnontouch user interface designgaze direction detectioncalibration-free Nomenclature UL=the gaze state estimated at the upper left directionUR=the gaze state estimated at the upper right directionDL=the gaze state estimated at the down left directionDR=the gaze state estimated1 at the down right directionC_pupil=the central coordinate position of the right or left pupilC_eye=the central coordinate position of the right or left eyeV_d=the difference vector between two central coordinate positionsX1=the central coordinate position of X-axis of the pupil’s bounding boxY1=the central coordinate position of Y-axis of the pupil’s bounding boxX2=the central coordinate position of X-axis of the eye’s bounding boxY2=the central coordinate position of Y-axis of the eye’s bounding boxTN=the number of true negative casesTP=the number of true positive casesFN=the number of false negative casesFP=the number of false positive casesF1 Score=it is a measure of a test’s accuracy by using 2×Precision×Recall/(Precision + Recall)mAP=it is a metric used to measure the performance of models doing object detection tasksDisclosure statementNo potential conflict of interest was reported by the author(s).Additional informationFundingThis work was financially supported by the Ministry of Science and Technology (MOST) under Grant No. [109-2218-E-005-008].
基于YOLOv3-tiny模型的智能广告显示客户凝视方向检测技术
摘要由于2019冠状病毒病(COVID-19)的流行,用于广告显示器或自动售货机的非触摸用户界面设计的凝视跟踪已成为一个新兴的研究课题。在本研究中,开发了一种基于深度学习的具有成本效益的智能广告显示客户注视方向检测技术。为了实现客户和显示器之间无需校准的交互,使用基于you - only - lookonce (YOLO)-v3-tiny的深度学习模型来确定眼睛和瞳孔的边界框。然后,利用投票机制和边界框中心坐标之间的差向量进行后处理,有效预测顾客凝视方向。产品图像被分成两个或四个凝视区。在跨人测试中,两个注视区域的查全率、查准率、查准率和f1得分分别约为77%、99%、88%和87%,四个注视区域的查全率、查准率和f1得分分别约为72%、91%、91%和79%。在NVIDIA图形处理单元加速嵌入式平台上的软件实现显示出接近每秒30帧的帧速率。该设计实现了智能广告平台的实时注视方向检测。副主编:袁淑明深度学习yolov3 - tiny智能系统智能显示器非触控用户界面设计凝视方向检测免校准术语UL=左上方向估计的凝视状态ur =右上方向估计的凝视状态dl =左下方向估计的凝视状态dr =右下方向估计的凝视状态c_瞳孔=右或左瞳孔的中心坐标位置c_eye =右或左眼睛的中心坐标位置v_d =两个中心坐标差向量x1 =瞳孔边界框的x轴中心坐标位置xy1 =瞳孔边界框的y轴中心坐标位置xx2 =眼睛边界框的x轴中心坐标位置xy2 =眼睛边界框的y轴中心坐标位置tn =真阴性病例数estp =真阳性病例数fn =假阴性病例数fp =假阳性病例数f1 Score=it是通过使用2×Precision×Recall/(Precision + Recall)mAP=它是用来衡量模型执行对象检测任务的性能的度量披露声明作者没有报告潜在的利益冲突。本研究由国家科技部(科技部)资助,批准号:(109 - 2218 - e - 005 - 008]。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Journal of the Chinese Institute of Engineers
Journal of the Chinese Institute of Engineers 工程技术-工程:综合
CiteScore
2.30
自引率
9.10%
发文量
57
审稿时长
6.8 months
期刊介绍: Encompassing a wide range of engineering disciplines and industrial applications, JCIE includes the following topics: 1.Chemical engineering 2.Civil engineering 3.Computer engineering 4.Electrical engineering 5.Electronics 6.Mechanical engineering and fields related to the above.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信