{"title":"RTL-Net: real-time lightweight Urban traffic object detection algorithm","authors":"Zhiqing Cui, Jiahao Yuan, Haibin Xu, Yamei Wei, Zhenglong Ding","doi":"10.1007/s40747-025-01875-z","DOIUrl":null,"url":null,"abstract":"<p>Object detection algorithm in urban traffic using remote sensing images often suffers from high complexity, low real-time performance, and low accuracy. To address these challenges, we propose RTL-Net, an urban traffic object detection network structure based on You Only Look Once (YOLO) v8s. To enhance real-time performance beyond the benchmark, we implemented lightweight designs for the loss function, backbone, neck, and head components. Firstly, a Powerable-IoU (PIoU) loss function was introduced to make the algorithm more suitable for different scales of targets and reduce false detection. Secondly, the Lightweight Shared Convolutional Detection (LSCD) head was replaced to ensure the detection performance and significantly improve the lightweight performance of the algorithm. Additionally, this paper introduces the Dilatation-wise Residual (DWR) module to facilitate the algorithm’s extraction of detailed features. In addition, we optimize the Bidirectional Feature Pyramid Network (Bi-FPN), enabling the fusion of multiple features to improve overall feature integration and performance. The VisDrone2021 dataset was utilized for experimental training. Experimental results demonstrate that the proposed algorithm achieves a significant 43.9% reduction in parameters and an 18.9% decrease in computational complexity. Moreover, the detection accuracy has improved by 2.3%, while maintaining a real-time detection speed of 263.2 frames per second. For edge computing object detection, our method outperforms YOLOv8s and leading remote sensing algorithms in both speed and accuracy, achieving state-of-the-art performance among single-stage detectors with comparable parameters.</p>","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"45 1","pages":""},"PeriodicalIF":4.6000,"publicationDate":"2025-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Complex & Intelligent Systems","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s40747-025-01875-z","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Object detection algorithm in urban traffic using remote sensing images often suffers from high complexity, low real-time performance, and low accuracy. To address these challenges, we propose RTL-Net, an urban traffic object detection network structure based on You Only Look Once (YOLO) v8s. To enhance real-time performance beyond the benchmark, we implemented lightweight designs for the loss function, backbone, neck, and head components. Firstly, a Powerable-IoU (PIoU) loss function was introduced to make the algorithm more suitable for different scales of targets and reduce false detection. Secondly, the Lightweight Shared Convolutional Detection (LSCD) head was replaced to ensure the detection performance and significantly improve the lightweight performance of the algorithm. Additionally, this paper introduces the Dilatation-wise Residual (DWR) module to facilitate the algorithm’s extraction of detailed features. In addition, we optimize the Bidirectional Feature Pyramid Network (Bi-FPN), enabling the fusion of multiple features to improve overall feature integration and performance. The VisDrone2021 dataset was utilized for experimental training. Experimental results demonstrate that the proposed algorithm achieves a significant 43.9% reduction in parameters and an 18.9% decrease in computational complexity. Moreover, the detection accuracy has improved by 2.3%, while maintaining a real-time detection speed of 263.2 frames per second. For edge computing object detection, our method outperforms YOLOv8s and leading remote sensing algorithms in both speed and accuracy, achieving state-of-the-art performance among single-stage detectors with comparable parameters.
基于遥感影像的城市交通目标检测算法往往存在复杂度高、实时性差、精度低等问题。为了应对这些挑战,我们提出了基于You Only Look Once (YOLO) v8s的城市交通目标检测网络结构RTL-Net。为了提高实时性能,我们对损失函数、主干、颈部和头部组件实施了轻量级设计。首先,引入power - iou (PIoU)损失函数,使算法更适合不同尺度的目标,减少误检;其次,替换轻量级共享卷积检测(LSCD)头,保证检测性能,显著提高算法的轻量级性能。此外,本文还引入了扩展残差(DWR)模块,以方便算法对细节特征的提取。此外,我们还优化了双向特征金字塔网络(Bi-FPN),实现了多个特征的融合,从而提高了整体特征的集成度和性能。使用VisDrone2021数据集进行实验训练。实验结果表明,该算法的参数减少43.9%,计算复杂度降低18.9%。检测精度提高2.3%,同时保持263.2帧/秒的实时检测速度。对于边缘计算目标检测,我们的方法在速度和精度上都优于YOLOv8s和领先的遥感算法,在具有可比参数的单级探测器中实现了最先进的性能。
期刊介绍:
Complex & Intelligent Systems aims to provide a forum for presenting and discussing novel approaches, tools and techniques meant for attaining a cross-fertilization between the broad fields of complex systems, computational simulation, and intelligent analytics and visualization. The transdisciplinary research that the journal focuses on will expand the boundaries of our understanding by investigating the principles and processes that underlie many of the most profound problems facing society today.