{"title":"A modeled study of driver visual attention driven by driving tasks","authors":"Chuan Xu , Bo Jiang , Yukun Wang , Yan Su","doi":"10.1016/j.engappai.2025.111382","DOIUrl":null,"url":null,"abstract":"<div><div>Visual attention is an indispensable component of driving, enabling drivers to swiftly identify critical objects within complex and dynamic traffic environments. Despite its significance, existing visual attention models predominantly focus on static or idealized driving scenarios, limiting their ability to capture attention distribution patterns in real-world, dynamic environments. Furthermore, most of these models rely heavily on data-driven approaches, extracting features exclusively from visual image data, while neglecting the profound influence of “the driver, the vehicle, and the road environment”. Consequently, these models frequently fail to effectively address the intricacies of practical driving scenarios. To bridge these gaps, this study introduces a driver visual attention prediction model that comprehensively incorporates the driving task, driver experience, and the impact of dynamic visual scenes. The proposed model leverages the advanced learning capabilities of Convolutional Neural Networks (CNN) and Vision Transformer (ViT), coupled with sequence modeling mechanisms, to effectively capture the nuanced attention allocation patterns of drivers in complex driving contexts. The model is meticulously designed to adapt to dynamically evolving driving task requirements. Experimental results demonstrate that the proposed model outperforms state-of-the-art (SOTA) visual attention prediction models across multiple benchmark evaluation metrics on the DR(eye)VE dataset, particularly excelling in dynamic driving conditions. Moreover, generalization experiments were conducted on the BDD-A and TDV datasets validate the model’s robustness and applicability across varied driving tasks and dynamic conditions.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"157 ","pages":"Article 111382"},"PeriodicalIF":7.5000,"publicationDate":"2025-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0952197625013843","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Visual attention is an indispensable component of driving, enabling drivers to swiftly identify critical objects within complex and dynamic traffic environments. Despite its significance, existing visual attention models predominantly focus on static or idealized driving scenarios, limiting their ability to capture attention distribution patterns in real-world, dynamic environments. Furthermore, most of these models rely heavily on data-driven approaches, extracting features exclusively from visual image data, while neglecting the profound influence of “the driver, the vehicle, and the road environment”. Consequently, these models frequently fail to effectively address the intricacies of practical driving scenarios. To bridge these gaps, this study introduces a driver visual attention prediction model that comprehensively incorporates the driving task, driver experience, and the impact of dynamic visual scenes. The proposed model leverages the advanced learning capabilities of Convolutional Neural Networks (CNN) and Vision Transformer (ViT), coupled with sequence modeling mechanisms, to effectively capture the nuanced attention allocation patterns of drivers in complex driving contexts. The model is meticulously designed to adapt to dynamically evolving driving task requirements. Experimental results demonstrate that the proposed model outperforms state-of-the-art (SOTA) visual attention prediction models across multiple benchmark evaluation metrics on the DR(eye)VE dataset, particularly excelling in dynamic driving conditions. Moreover, generalization experiments were conducted on the BDD-A and TDV datasets validate the model’s robustness and applicability across varied driving tasks and dynamic conditions.
期刊介绍:
Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.