Danfeng Du , Mengju Bi , Yuchen Xie , Yang Liu , Guanlin Qi , Yangyang Guo
{"title":"MLE-YOLO:用于自动驾驶恶劣天气的轻型、坚固的车辆和行人探测器","authors":"Danfeng Du , Mengju Bi , Yuchen Xie , Yang Liu , Guanlin Qi , Yangyang Guo","doi":"10.1016/j.dsp.2025.105628","DOIUrl":null,"url":null,"abstract":"<div><div>Adverse weather poses significant challenges to object detection in autonomous driving, including poor visibility, precipitation interference, and motion blur. Additionally, conventional object detectors often struggle to balance computational efficiency with detection accuracy in such conditions. To address these issues, we propose MLE-YOLO (Multimodal Lightweight Enhanced YOLO), a multimodal fusion framework built upon YOLOv11, optimized for robust vehicle and pedestrian detection in adverse environments. MLE-YOLO integrates four key innovations. The backbone is enhanced with a Multi-Stage Partial Transformer Module (M-SPTM), which combines CNN and Transformer branches to improve detection accuracy in foggy and rainy scenarios while maintaining computational efficiency. The neck adopts a Mixed Aggregation Network (MANet) that leverages depthwise separable convolutions and dynamic cross-layer connections to strengthen multi-scale feature fusion and suppress weather-induced noise. A Lightweight Downsampling Module (LDM) is designed to enhance small-object detection through multi-path feature aggregation, coupled with a compact structure that reduces computational load. Finally, an Efficient Lightweight Detection Head (ELDH) incorporates detail-enhancing convolutions to extract both intensity and gradient features from degraded visual inputs. Extensive experiments on a custom adverse weather dataset demonstrate that MLE-YOLO improves F1 score by 3.6 % and mAP by 3.0 %, while reducing model parameters by 15.8 %, model size by 7.1 %, and FLOPs by 3.2 %. These results validate MLE-YOLO as a lightweight and robust solution for real-time perception in autonomous driving under challenging environmental conditions.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"168 ","pages":"Article 105628"},"PeriodicalIF":3.0000,"publicationDate":"2025-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"MLE-YOLO: A lightweight and robust vehicle and pedestrian detector for adverse weather in autonomous driving\",\"authors\":\"Danfeng Du , Mengju Bi , Yuchen Xie , Yang Liu , Guanlin Qi , Yangyang Guo\",\"doi\":\"10.1016/j.dsp.2025.105628\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Adverse weather poses significant challenges to object detection in autonomous driving, including poor visibility, precipitation interference, and motion blur. Additionally, conventional object detectors often struggle to balance computational efficiency with detection accuracy in such conditions. To address these issues, we propose MLE-YOLO (Multimodal Lightweight Enhanced YOLO), a multimodal fusion framework built upon YOLOv11, optimized for robust vehicle and pedestrian detection in adverse environments. MLE-YOLO integrates four key innovations. The backbone is enhanced with a Multi-Stage Partial Transformer Module (M-SPTM), which combines CNN and Transformer branches to improve detection accuracy in foggy and rainy scenarios while maintaining computational efficiency. The neck adopts a Mixed Aggregation Network (MANet) that leverages depthwise separable convolutions and dynamic cross-layer connections to strengthen multi-scale feature fusion and suppress weather-induced noise. A Lightweight Downsampling Module (LDM) is designed to enhance small-object detection through multi-path feature aggregation, coupled with a compact structure that reduces computational load. Finally, an Efficient Lightweight Detection Head (ELDH) incorporates detail-enhancing convolutions to extract both intensity and gradient features from degraded visual inputs. Extensive experiments on a custom adverse weather dataset demonstrate that MLE-YOLO improves F1 score by 3.6 % and mAP by 3.0 %, while reducing model parameters by 15.8 %, model size by 7.1 %, and FLOPs by 3.2 %. These results validate MLE-YOLO as a lightweight and robust solution for real-time perception in autonomous driving under challenging environmental conditions.</div></div>\",\"PeriodicalId\":51011,\"journal\":{\"name\":\"Digital Signal Processing\",\"volume\":\"168 \",\"pages\":\"Article 105628\"},\"PeriodicalIF\":3.0000,\"publicationDate\":\"2025-09-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Digital Signal Processing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1051200425006505\",\"RegionNum\":3,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Digital Signal Processing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1051200425006505","RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
MLE-YOLO: A lightweight and robust vehicle and pedestrian detector for adverse weather in autonomous driving
Adverse weather poses significant challenges to object detection in autonomous driving, including poor visibility, precipitation interference, and motion blur. Additionally, conventional object detectors often struggle to balance computational efficiency with detection accuracy in such conditions. To address these issues, we propose MLE-YOLO (Multimodal Lightweight Enhanced YOLO), a multimodal fusion framework built upon YOLOv11, optimized for robust vehicle and pedestrian detection in adverse environments. MLE-YOLO integrates four key innovations. The backbone is enhanced with a Multi-Stage Partial Transformer Module (M-SPTM), which combines CNN and Transformer branches to improve detection accuracy in foggy and rainy scenarios while maintaining computational efficiency. The neck adopts a Mixed Aggregation Network (MANet) that leverages depthwise separable convolutions and dynamic cross-layer connections to strengthen multi-scale feature fusion and suppress weather-induced noise. A Lightweight Downsampling Module (LDM) is designed to enhance small-object detection through multi-path feature aggregation, coupled with a compact structure that reduces computational load. Finally, an Efficient Lightweight Detection Head (ELDH) incorporates detail-enhancing convolutions to extract both intensity and gradient features from degraded visual inputs. Extensive experiments on a custom adverse weather dataset demonstrate that MLE-YOLO improves F1 score by 3.6 % and mAP by 3.0 %, while reducing model parameters by 15.8 %, model size by 7.1 %, and FLOPs by 3.2 %. These results validate MLE-YOLO as a lightweight and robust solution for real-time perception in autonomous driving under challenging environmental conditions.
期刊介绍:
Digital Signal Processing: A Review Journal is one of the oldest and most established journals in the field of signal processing yet it aims to be the most innovative. The Journal invites top quality research articles at the frontiers of research in all aspects of signal processing. Our objective is to provide a platform for the publication of ground-breaking research in signal processing with both academic and industrial appeal.
The journal has a special emphasis on statistical signal processing methodology such as Bayesian signal processing, and encourages articles on emerging applications of signal processing such as:
• big data• machine learning• internet of things• information security• systems biology and computational biology,• financial time series analysis,• autonomous vehicles,• quantum computing,• neuromorphic engineering,• human-computer interaction and intelligent user interfaces,• environmental signal processing,• geophysical signal processing including seismic signal processing,• chemioinformatics and bioinformatics,• audio, visual and performance arts,• disaster management and prevention,• renewable energy,