{"title":"Convolutional Dual-Attention-Network (CDAN): A multiple light intensities based driver emotion recognition method","authors":"Ahad Ahamed , Xiaohui Yang , Tao Xu , Qingbei Guo","doi":"10.1016/j.jvcir.2025.104558","DOIUrl":null,"url":null,"abstract":"<div><div>Driver emotion recognition is critical for enhancing traffic safety and influencing driver behavior. However, current methods struggle to accurately classify emotions under variable lighting conditions such as bright sunlight, shadows, and low light environments, resulting in inconsistent feature extraction and reduced accuracy. Moreover, many approaches incur high computational costs and excessive feature exchanges, limiting real-world deployment in resource-constrained settings. To address these challenges, we propose the Convolutional Dual-Attention Network (CDAN), a novel framework designed to mitigate the impact of light intensity variations in driving scenarios. Our framework integrates Multi-Convolutional Linear Layer Attention (MCLLA), which leverages linear attention augmented with Rotary Positional Encoding (RoPE) and Locally Enhanced Positional Encoding (LePE) to capture global and local spatial relationships. Additionally, a Convolutional Attention Module (CAM) refines feature maps to improve representation quality. Evaluations of MLI-DER, modified KMU-FED, and CK+ datasets demonstrate its enhanced effectiveness compared to existing methods in handling diverse lighting conditions.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"112 ","pages":"Article 104558"},"PeriodicalIF":3.1000,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Visual Communication and Image Representation","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1047320325001725","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Driver emotion recognition is critical for enhancing traffic safety and influencing driver behavior. However, current methods struggle to accurately classify emotions under variable lighting conditions such as bright sunlight, shadows, and low light environments, resulting in inconsistent feature extraction and reduced accuracy. Moreover, many approaches incur high computational costs and excessive feature exchanges, limiting real-world deployment in resource-constrained settings. To address these challenges, we propose the Convolutional Dual-Attention Network (CDAN), a novel framework designed to mitigate the impact of light intensity variations in driving scenarios. Our framework integrates Multi-Convolutional Linear Layer Attention (MCLLA), which leverages linear attention augmented with Rotary Positional Encoding (RoPE) and Locally Enhanced Positional Encoding (LePE) to capture global and local spatial relationships. Additionally, a Convolutional Attention Module (CAM) refines feature maps to improve representation quality. Evaluations of MLI-DER, modified KMU-FED, and CK+ datasets demonstrate its enhanced effectiveness compared to existing methods in handling diverse lighting conditions.
期刊介绍:
The Journal of Visual Communication and Image Representation publishes papers on state-of-the-art visual communication and image representation, with emphasis on novel technologies and theoretical work in this multidisciplinary area of pure and applied research. The field of visual communication and image representation is considered in its broadest sense and covers both digital and analog aspects as well as processing and communication in biological visual systems.