Feng You , Yi Xie , Siyi Zhang , Hao Chen , Haiwei Wang , Wei Zhang , Jianrong Liu
{"title":"Attention based network for real-time road drivable area, lane line detection and scene identification","authors":"Feng You , Yi Xie , Siyi Zhang , Hao Chen , Haiwei Wang , Wei Zhang , Jianrong Liu","doi":"10.1016/j.engappai.2025.111781","DOIUrl":null,"url":null,"abstract":"<div><div>The detection of road drivable areas and lane lines is considered a fundamental component of autonomous driving systems. However, most existing approaches handle these tasks independently, and multi-task networks frequently neglect the inherent correlation between them while failing to differentiate various lane line types. In practice, the delineation of drivable regions is strongly influenced by both lane line characteristics and contextual street scenes. To address these limitations, a novel multi-task network—Real-time Road Drivable Area, Lane Line Detection, and Scene Identification Network (RLSNet)—is proposed. This network is designed to perform simultaneous segmentation of drivable areas, detection of lane lines, and classification of road scenes. Drivable area estimation is optimized through the integration of lane and scene cues, guided by traffic regulations. A Residual Network (ResNet)-based backbone is employed, enhanced with Bidirectional Fusion Attention (BFA) for feature encoding. This is followed by a decoder incorporating a Feature Aggregation Module (FAM) to enable effective semantic–spatial fusion. Lane line detection is further refined using a Bilateral Up-Sampling Decoder (BUSD), while scene understanding is enhanced via a Scene Classification Module (SCM). Extensive experiments conducted on the challenging Berkeley DeepDrive 100K(BDD100K) dataset have demonstrated that RLSNet achieves high accuracy in both drivable area and lane line detection by leveraging the mutual guidance of lane and scene information. Furthermore, the network maintains real-time inference speed at 93 frames per second (FPS), striking a practical balance between semantic fidelity and computational efficiency for real-world deployment. The implementation code has been made publicly available at: <span><span>https://github.com/033186ZSY/RLSNet-master</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"160 ","pages":"Article 111781"},"PeriodicalIF":7.5000,"publicationDate":"2025-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S095219762501783X","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
The detection of road drivable areas and lane lines is considered a fundamental component of autonomous driving systems. However, most existing approaches handle these tasks independently, and multi-task networks frequently neglect the inherent correlation between them while failing to differentiate various lane line types. In practice, the delineation of drivable regions is strongly influenced by both lane line characteristics and contextual street scenes. To address these limitations, a novel multi-task network—Real-time Road Drivable Area, Lane Line Detection, and Scene Identification Network (RLSNet)—is proposed. This network is designed to perform simultaneous segmentation of drivable areas, detection of lane lines, and classification of road scenes. Drivable area estimation is optimized through the integration of lane and scene cues, guided by traffic regulations. A Residual Network (ResNet)-based backbone is employed, enhanced with Bidirectional Fusion Attention (BFA) for feature encoding. This is followed by a decoder incorporating a Feature Aggregation Module (FAM) to enable effective semantic–spatial fusion. Lane line detection is further refined using a Bilateral Up-Sampling Decoder (BUSD), while scene understanding is enhanced via a Scene Classification Module (SCM). Extensive experiments conducted on the challenging Berkeley DeepDrive 100K(BDD100K) dataset have demonstrated that RLSNet achieves high accuracy in both drivable area and lane line detection by leveraging the mutual guidance of lane and scene information. Furthermore, the network maintains real-time inference speed at 93 frames per second (FPS), striking a practical balance between semantic fidelity and computational efficiency for real-world deployment. The implementation code has been made publicly available at: https://github.com/033186ZSY/RLSNet-master.
期刊介绍:
Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.