MLE-YOLO:用于自动驾驶恶劣天气的轻型、坚固的车辆和行人探测器

IF 3 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC
Danfeng Du , Mengju Bi , Yuchen Xie , Yang Liu , Guanlin Qi , Yangyang Guo
{"title":"MLE-YOLO:用于自动驾驶恶劣天气的轻型、坚固的车辆和行人探测器","authors":"Danfeng Du ,&nbsp;Mengju Bi ,&nbsp;Yuchen Xie ,&nbsp;Yang Liu ,&nbsp;Guanlin Qi ,&nbsp;Yangyang Guo","doi":"10.1016/j.dsp.2025.105628","DOIUrl":null,"url":null,"abstract":"<div><div>Adverse weather poses significant challenges to object detection in autonomous driving, including poor visibility, precipitation interference, and motion blur. Additionally, conventional object detectors often struggle to balance computational efficiency with detection accuracy in such conditions. To address these issues, we propose MLE-YOLO (Multimodal Lightweight Enhanced YOLO), a multimodal fusion framework built upon YOLOv11, optimized for robust vehicle and pedestrian detection in adverse environments. MLE-YOLO integrates four key innovations. The backbone is enhanced with a Multi-Stage Partial Transformer Module (M-SPTM), which combines CNN and Transformer branches to improve detection accuracy in foggy and rainy scenarios while maintaining computational efficiency. The neck adopts a Mixed Aggregation Network (MANet) that leverages depthwise separable convolutions and dynamic cross-layer connections to strengthen multi-scale feature fusion and suppress weather-induced noise. A Lightweight Downsampling Module (LDM) is designed to enhance small-object detection through multi-path feature aggregation, coupled with a compact structure that reduces computational load. Finally, an Efficient Lightweight Detection Head (ELDH) incorporates detail-enhancing convolutions to extract both intensity and gradient features from degraded visual inputs. Extensive experiments on a custom adverse weather dataset demonstrate that MLE-YOLO improves F1 score by 3.6 % and mAP by 3.0 %, while reducing model parameters by 15.8 %, model size by 7.1 %, and FLOPs by 3.2 %. These results validate MLE-YOLO as a lightweight and robust solution for real-time perception in autonomous driving under challenging environmental conditions.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"168 ","pages":"Article 105628"},"PeriodicalIF":3.0000,"publicationDate":"2025-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"MLE-YOLO: A lightweight and robust vehicle and pedestrian detector for adverse weather in autonomous driving\",\"authors\":\"Danfeng Du ,&nbsp;Mengju Bi ,&nbsp;Yuchen Xie ,&nbsp;Yang Liu ,&nbsp;Guanlin Qi ,&nbsp;Yangyang Guo\",\"doi\":\"10.1016/j.dsp.2025.105628\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Adverse weather poses significant challenges to object detection in autonomous driving, including poor visibility, precipitation interference, and motion blur. Additionally, conventional object detectors often struggle to balance computational efficiency with detection accuracy in such conditions. To address these issues, we propose MLE-YOLO (Multimodal Lightweight Enhanced YOLO), a multimodal fusion framework built upon YOLOv11, optimized for robust vehicle and pedestrian detection in adverse environments. MLE-YOLO integrates four key innovations. The backbone is enhanced with a Multi-Stage Partial Transformer Module (M-SPTM), which combines CNN and Transformer branches to improve detection accuracy in foggy and rainy scenarios while maintaining computational efficiency. The neck adopts a Mixed Aggregation Network (MANet) that leverages depthwise separable convolutions and dynamic cross-layer connections to strengthen multi-scale feature fusion and suppress weather-induced noise. A Lightweight Downsampling Module (LDM) is designed to enhance small-object detection through multi-path feature aggregation, coupled with a compact structure that reduces computational load. Finally, an Efficient Lightweight Detection Head (ELDH) incorporates detail-enhancing convolutions to extract both intensity and gradient features from degraded visual inputs. Extensive experiments on a custom adverse weather dataset demonstrate that MLE-YOLO improves F1 score by 3.6 % and mAP by 3.0 %, while reducing model parameters by 15.8 %, model size by 7.1 %, and FLOPs by 3.2 %. These results validate MLE-YOLO as a lightweight and robust solution for real-time perception in autonomous driving under challenging environmental conditions.</div></div>\",\"PeriodicalId\":51011,\"journal\":{\"name\":\"Digital Signal Processing\",\"volume\":\"168 \",\"pages\":\"Article 105628\"},\"PeriodicalIF\":3.0000,\"publicationDate\":\"2025-09-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Digital Signal Processing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1051200425006505\",\"RegionNum\":3,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Digital Signal Processing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1051200425006505","RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

摘要

恶劣天气给自动驾驶中的目标检测带来了重大挑战,包括能见度低、降水干扰和运动模糊。此外,在这种情况下,传统的目标检测器通常难以平衡计算效率和检测精度。为了解决这些问题,我们提出了MLE-YOLO (Multimodal Lightweight Enhanced YOLO),这是一种基于YOLOv11的多模式融合框架,针对恶劣环境下的车辆和行人检测进行了优化。MLE-YOLO集成了四个关键创新。主干网采用多级局部变压器模块(M-SPTM)进行增强,该模块结合了CNN和Transformer分支,以提高雾天和雨天场景下的检测精度,同时保持计算效率。颈部采用混合聚合网络(MANet),利用深度可分卷积和动态跨层连接来加强多尺度特征融合,抑制天气引起的噪声。轻量级下采样模块(LDM)旨在通过多路径特征聚合来增强小目标检测,并结合紧凑的结构来减少计算负载。最后,高效轻量级检测头(ELDH)结合了细节增强卷积,从退化的视觉输入中提取强度和梯度特征。在自定义恶劣天气数据集上进行的大量实验表明,MLE-YOLO将F1分数提高了3.6%,mAP提高了3.0%,同时将模型参数降低了15.8%,模型大小降低了7.1%,FLOPs降低了3.2%。这些结果验证了MLE-YOLO是一种轻量级且强大的解决方案,可在具有挑战性的环境条件下用于自动驾驶的实时感知。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
MLE-YOLO: A lightweight and robust vehicle and pedestrian detector for adverse weather in autonomous driving
Adverse weather poses significant challenges to object detection in autonomous driving, including poor visibility, precipitation interference, and motion blur. Additionally, conventional object detectors often struggle to balance computational efficiency with detection accuracy in such conditions. To address these issues, we propose MLE-YOLO (Multimodal Lightweight Enhanced YOLO), a multimodal fusion framework built upon YOLOv11, optimized for robust vehicle and pedestrian detection in adverse environments. MLE-YOLO integrates four key innovations. The backbone is enhanced with a Multi-Stage Partial Transformer Module (M-SPTM), which combines CNN and Transformer branches to improve detection accuracy in foggy and rainy scenarios while maintaining computational efficiency. The neck adopts a Mixed Aggregation Network (MANet) that leverages depthwise separable convolutions and dynamic cross-layer connections to strengthen multi-scale feature fusion and suppress weather-induced noise. A Lightweight Downsampling Module (LDM) is designed to enhance small-object detection through multi-path feature aggregation, coupled with a compact structure that reduces computational load. Finally, an Efficient Lightweight Detection Head (ELDH) incorporates detail-enhancing convolutions to extract both intensity and gradient features from degraded visual inputs. Extensive experiments on a custom adverse weather dataset demonstrate that MLE-YOLO improves F1 score by 3.6 % and mAP by 3.0 %, while reducing model parameters by 15.8 %, model size by 7.1 %, and FLOPs by 3.2 %. These results validate MLE-YOLO as a lightweight and robust solution for real-time perception in autonomous driving under challenging environmental conditions.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Digital Signal Processing
Digital Signal Processing 工程技术-工程:电子与电气
CiteScore
5.30
自引率
17.20%
发文量
435
审稿时长
66 days
期刊介绍: Digital Signal Processing: A Review Journal is one of the oldest and most established journals in the field of signal processing yet it aims to be the most innovative. The Journal invites top quality research articles at the frontiers of research in all aspects of signal processing. Our objective is to provide a platform for the publication of ground-breaking research in signal processing with both academic and industrial appeal. The journal has a special emphasis on statistical signal processing methodology such as Bayesian signal processing, and encourages articles on emerging applications of signal processing such as: • big data• machine learning• internet of things• information security• systems biology and computational biology,• financial time series analysis,• autonomous vehicles,• quantum computing,• neuromorphic engineering,• human-computer interaction and intelligent user interfaces,• environmental signal processing,• geophysical signal processing including seismic signal processing,• chemioinformatics and bioinformatics,• audio, visual and performance arts,• disaster management and prevention,• renewable energy,
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信