基于时空双分支特征引导的驾驶员注意力预测融合网络

IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Yuekui Zhang , Yunzuo Zhang , Yaoge Xiao , Tong Wang
{"title":"基于时空双分支特征引导的驾驶员注意力预测融合网络","authors":"Yuekui Zhang ,&nbsp;Yunzuo Zhang ,&nbsp;Yaoge Xiao ,&nbsp;Tong Wang","doi":"10.1016/j.eswa.2025.128564","DOIUrl":null,"url":null,"abstract":"<div><div>Predicting the driver’s gaze area is crucial for safe driving in rapidly changing traffic scenarios. However, existing driver attention prediction models generally suffer from two key limitations: insufficient utilization of spatial scale features, which hinders the precise capture of critical information in the scene; the lack of effective guidance from motion information between video frames, making it difficult to assess dynamic changes in the surrounding environment accurately. To address these issues, we propose a Spatiotemporal Dual-branch Feature-guided Fusion Network (SDFF-Net). Specifically, in the spatial branch, we design a Multi-scale Feature Aggregation (MFA) module to enhance the representation of detailed features by constructing bidirectional sampling and layer-by-layer correlation paths, enabling comprehensive extraction of saliency cues across receptive fields. In the temporal branch, we introduce an Attention Transfer Mechanism (ATM) to guide temporal modeling across consecutive frames, improving the ability to capture long-distance dependencies. Finally, we fuse the spatiotemporal features and decode them to generate the predicted saliency map. Experimental results on the DADA-2000 and TDV datasets show that the proposed SDFF-Net achieves state-of-the-art performance in driver attention prediction, outperforming existing methods in multiple evaluation metrics. Benefiting from its efficient dual-branch architecture, SDFF-Net is well-suited for deployment in resource-constrained environments, providing reliable real-time attention prediction, which is of great significance for enhancing driving safety and supporting advanced driver assistance systems in complex traffic scenarios.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"292 ","pages":"Article 128564"},"PeriodicalIF":7.5000,"publicationDate":"2025-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Spatiotemporal dual-branch feature-guided fusion network for driver attention prediction\",\"authors\":\"Yuekui Zhang ,&nbsp;Yunzuo Zhang ,&nbsp;Yaoge Xiao ,&nbsp;Tong Wang\",\"doi\":\"10.1016/j.eswa.2025.128564\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Predicting the driver’s gaze area is crucial for safe driving in rapidly changing traffic scenarios. However, existing driver attention prediction models generally suffer from two key limitations: insufficient utilization of spatial scale features, which hinders the precise capture of critical information in the scene; the lack of effective guidance from motion information between video frames, making it difficult to assess dynamic changes in the surrounding environment accurately. To address these issues, we propose a Spatiotemporal Dual-branch Feature-guided Fusion Network (SDFF-Net). Specifically, in the spatial branch, we design a Multi-scale Feature Aggregation (MFA) module to enhance the representation of detailed features by constructing bidirectional sampling and layer-by-layer correlation paths, enabling comprehensive extraction of saliency cues across receptive fields. In the temporal branch, we introduce an Attention Transfer Mechanism (ATM) to guide temporal modeling across consecutive frames, improving the ability to capture long-distance dependencies. Finally, we fuse the spatiotemporal features and decode them to generate the predicted saliency map. Experimental results on the DADA-2000 and TDV datasets show that the proposed SDFF-Net achieves state-of-the-art performance in driver attention prediction, outperforming existing methods in multiple evaluation metrics. Benefiting from its efficient dual-branch architecture, SDFF-Net is well-suited for deployment in resource-constrained environments, providing reliable real-time attention prediction, which is of great significance for enhancing driving safety and supporting advanced driver assistance systems in complex traffic scenarios.</div></div>\",\"PeriodicalId\":50461,\"journal\":{\"name\":\"Expert Systems with Applications\",\"volume\":\"292 \",\"pages\":\"Article 128564\"},\"PeriodicalIF\":7.5000,\"publicationDate\":\"2025-06-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Expert Systems with Applications\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0957417425021839\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Expert Systems with Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0957417425021839","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

摘要

在快速变化的交通场景中,预测驾驶员的视线区域对于安全驾驶至关重要。然而,现有的驾驶员注意力预测模型普遍存在两个关键的局限性:对空间尺度特征的利用不足,阻碍了对场景中关键信息的精确捕获;缺乏视频帧之间运动信息的有效引导,难以准确评估周围环境的动态变化。为了解决这些问题,我们提出了一种时空双分支特征引导融合网络(SDFF-Net)。具体而言,在空间分支中,我们设计了一个多尺度特征聚合(MFA)模块,通过构建双向采样和逐层相关路径来增强细节特征的表示,从而实现跨感受场的显著性线索的综合提取。在时间分支中,我们引入了一种注意力转移机制(ATM)来指导跨连续帧的时间建模,提高了捕获远程依赖关系的能力。最后,我们将时空特征融合并解码,生成预测的显著性图。在DADA-2000和TDV数据集上的实验结果表明,所提出的SDFF-Net在驾驶员注意力预测方面达到了最先进的性能,在多个评估指标上优于现有方法。得益于其高效的双分支架构,SDFF-Net非常适合在资源受限环境下部署,提供可靠的实时注意力预测,对于提高驾驶安全性和支持复杂交通场景下的先进驾驶辅助系统具有重要意义。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Spatiotemporal dual-branch feature-guided fusion network for driver attention prediction
Predicting the driver’s gaze area is crucial for safe driving in rapidly changing traffic scenarios. However, existing driver attention prediction models generally suffer from two key limitations: insufficient utilization of spatial scale features, which hinders the precise capture of critical information in the scene; the lack of effective guidance from motion information between video frames, making it difficult to assess dynamic changes in the surrounding environment accurately. To address these issues, we propose a Spatiotemporal Dual-branch Feature-guided Fusion Network (SDFF-Net). Specifically, in the spatial branch, we design a Multi-scale Feature Aggregation (MFA) module to enhance the representation of detailed features by constructing bidirectional sampling and layer-by-layer correlation paths, enabling comprehensive extraction of saliency cues across receptive fields. In the temporal branch, we introduce an Attention Transfer Mechanism (ATM) to guide temporal modeling across consecutive frames, improving the ability to capture long-distance dependencies. Finally, we fuse the spatiotemporal features and decode them to generate the predicted saliency map. Experimental results on the DADA-2000 and TDV datasets show that the proposed SDFF-Net achieves state-of-the-art performance in driver attention prediction, outperforming existing methods in multiple evaluation metrics. Benefiting from its efficient dual-branch architecture, SDFF-Net is well-suited for deployment in resource-constrained environments, providing reliable real-time attention prediction, which is of great significance for enhancing driving safety and supporting advanced driver assistance systems in complex traffic scenarios.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Expert Systems with Applications
Expert Systems with Applications 工程技术-工程:电子与电气
CiteScore
13.80
自引率
10.60%
发文量
2045
审稿时长
8.7 months
期刊介绍: Expert Systems With Applications is an international journal dedicated to the exchange of information on expert and intelligent systems used globally in industry, government, and universities. The journal emphasizes original papers covering the design, development, testing, implementation, and management of these systems, offering practical guidelines. It spans various sectors such as finance, engineering, marketing, law, project management, information management, medicine, and more. The journal also welcomes papers on multi-agent systems, knowledge management, neural networks, knowledge discovery, data mining, and other related areas, excluding applications to military/defense systems.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信