Sayyed Mudassar Shah, Gan Zengkang, Zhaoyun Sun, Tariq Hussain, Khalid Zaman, Abdullah Alwabli, Amar Y. Jaffar, Farman Ali
{"title":"AI-enabled driver assistance: monitoring head and gaze movements for enhanced safety","authors":"Sayyed Mudassar Shah, Gan Zengkang, Zhaoyun Sun, Tariq Hussain, Khalid Zaman, Abdullah Alwabli, Amar Y. Jaffar, Farman Ali","doi":"10.1007/s40747-025-01897-7","DOIUrl":null,"url":null,"abstract":"<p>This paper introduces a real-time head-pose detection and eye-gaze estimation system for Automatic Driver Assistance Technology (ADAT) aimed at enhancing driver safety by accurately collecting and transmitting data on the driver’s head position and eye gaze to mitigate potential risks. Existing methods are constrained by significant limitations, including reduced accuracy under challenging conditions such as varying head orientations and lighting, higher latency in real-time applications (e.g., Faster-RCNN and TPH-YOLOv5), and computational inefficiency, which hinders their deployment in resource-constrained environments. To address these challenges, we propose a novel framework using the Transformer Detection of Gaze Head - YOLOv7 (TDGH-YOLOv7) object detector. The key contributions of this work include the development of a reference image dataset encompassing diverse vertical and horizontal gaze positions alongside the implementation of an optimized detection system that achieves state-of-the-art performance in terms of accuracy and latency. The proposed system achieves superior precision, with a weighted accuracy of 95.02% and Root Mean Square Errors of 2.23 and 1.68 for vertical and horizontal gaze estimation, respectively, validated on the MPII-Gaze and DG-Unicamp datasets. A comprehensive comparative analysis with existing models, such as CNN, SSD, Faster-RCNN, and YOLOv8, underscores the robustness and efficiency of the proposed approach. Finally, the implications of these findings are discussed, and potential avenues for future research are outlined.</p>","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"133 1","pages":""},"PeriodicalIF":5.0000,"publicationDate":"2025-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Complex & Intelligent Systems","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s40747-025-01897-7","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
This paper introduces a real-time head-pose detection and eye-gaze estimation system for Automatic Driver Assistance Technology (ADAT) aimed at enhancing driver safety by accurately collecting and transmitting data on the driver’s head position and eye gaze to mitigate potential risks. Existing methods are constrained by significant limitations, including reduced accuracy under challenging conditions such as varying head orientations and lighting, higher latency in real-time applications (e.g., Faster-RCNN and TPH-YOLOv5), and computational inefficiency, which hinders their deployment in resource-constrained environments. To address these challenges, we propose a novel framework using the Transformer Detection of Gaze Head - YOLOv7 (TDGH-YOLOv7) object detector. The key contributions of this work include the development of a reference image dataset encompassing diverse vertical and horizontal gaze positions alongside the implementation of an optimized detection system that achieves state-of-the-art performance in terms of accuracy and latency. The proposed system achieves superior precision, with a weighted accuracy of 95.02% and Root Mean Square Errors of 2.23 and 1.68 for vertical and horizontal gaze estimation, respectively, validated on the MPII-Gaze and DG-Unicamp datasets. A comprehensive comparative analysis with existing models, such as CNN, SSD, Faster-RCNN, and YOLOv8, underscores the robustness and efficiency of the proposed approach. Finally, the implications of these findings are discussed, and potential avenues for future research are outlined.
期刊介绍:
Complex & Intelligent Systems aims to provide a forum for presenting and discussing novel approaches, tools and techniques meant for attaining a cross-fertilization between the broad fields of complex systems, computational simulation, and intelligent analytics and visualization. The transdisciplinary research that the journal focuses on will expand the boundaries of our understanding by investigating the principles and processes that underlie many of the most profound problems facing society today.