{"title":"GLD-Road:一种面向遥感图像的全局-局部解码路网提取模型","authors":"Ligao Deng, Yupeng Deng, Yu Meng, Jingbo Chen, Zhihao Xi, Diyou Liu, Qifeng Chu","doi":"10.1016/j.isprsjprs.2025.07.026","DOIUrl":null,"url":null,"abstract":"Road networks are essential information for map updates, autonomous driving, and disaster response. However, manual annotation of road networks from remote sensing imagery is time-consuming and costly, whereas deep learning methods have gained attention for their efficiency and precision in road extraction. Current deep learning approaches for road network extraction fall into three main categories: postprocessing methods based on semantic segmentation results, global parallel methods and local iterative methods. Postprocessing methods introduce quantization errors, leading to higher overall road network inaccuracies; global parallel methods achieve high extraction efficiency but risk road node omissions; local iterative methods excel in node detection but have relatively lower extraction efficiency. To address the above limitations, We propose a two-stage road extraction model with global–local decoding, named GLD-Road, which possesses the high efficiency of global parallel methods and the strong node perception capability of local iterative methods, enabling a significant reduction in inference time while maintaining high-precision road network extraction. In the first stage, GLD-Road extracts the coordinates and direction descriptors of road nodes using global information from the entire input image. Subsequently, it connects adjacent nodes using a self-designed graph network module (Connect Module) to form the initial road network. In the second stage, based on the road endpoints contained in the initial road network, GLD-Road iteratively searches local images and the local grid map of the primary network to repair broken roads, ultimately producing a complete road network. Since the second stage only requires limited supplementary detection of locally missing nodes, GLD-Road significantly reduces the global iterative search range over the entire image, leading to a substantial reduction in retrieval time compared to local iterative methods. Finally, experimental results revealed that GLD-Road outperformed current state-of-the-art methods, achieving improvements of 1.9% and 0.67% in average path length similarity (APLS) on the City-Scale and SpaceNet3 datasets, respectively. Moreover, compared with those of a global parallel method (Sat2Graph) and a local iterative method (RNGDet++), the retrieval time of GLD-Road exhibited reductions of 40% and 92%, respectively, suggesting that GLD-Road achieves a pronounced improvement in road network extraction efficiency compared to existing methods. The experimental results are available at <ce:inter-ref xlink:href=\"https://github.com/ucas-dlg/GLD-Road\" xlink:type=\"simple\">https://github.com/ucas-dlg/GLD-Road</ce:inter-ref>.","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"31 1","pages":""},"PeriodicalIF":12.2000,"publicationDate":"2025-08-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"GLD-Road: A global–local decoding road network extraction model for remote sensing images\",\"authors\":\"Ligao Deng, Yupeng Deng, Yu Meng, Jingbo Chen, Zhihao Xi, Diyou Liu, Qifeng Chu\",\"doi\":\"10.1016/j.isprsjprs.2025.07.026\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Road networks are essential information for map updates, autonomous driving, and disaster response. However, manual annotation of road networks from remote sensing imagery is time-consuming and costly, whereas deep learning methods have gained attention for their efficiency and precision in road extraction. Current deep learning approaches for road network extraction fall into three main categories: postprocessing methods based on semantic segmentation results, global parallel methods and local iterative methods. Postprocessing methods introduce quantization errors, leading to higher overall road network inaccuracies; global parallel methods achieve high extraction efficiency but risk road node omissions; local iterative methods excel in node detection but have relatively lower extraction efficiency. To address the above limitations, We propose a two-stage road extraction model with global–local decoding, named GLD-Road, which possesses the high efficiency of global parallel methods and the strong node perception capability of local iterative methods, enabling a significant reduction in inference time while maintaining high-precision road network extraction. In the first stage, GLD-Road extracts the coordinates and direction descriptors of road nodes using global information from the entire input image. Subsequently, it connects adjacent nodes using a self-designed graph network module (Connect Module) to form the initial road network. In the second stage, based on the road endpoints contained in the initial road network, GLD-Road iteratively searches local images and the local grid map of the primary network to repair broken roads, ultimately producing a complete road network. Since the second stage only requires limited supplementary detection of locally missing nodes, GLD-Road significantly reduces the global iterative search range over the entire image, leading to a substantial reduction in retrieval time compared to local iterative methods. Finally, experimental results revealed that GLD-Road outperformed current state-of-the-art methods, achieving improvements of 1.9% and 0.67% in average path length similarity (APLS) on the City-Scale and SpaceNet3 datasets, respectively. Moreover, compared with those of a global parallel method (Sat2Graph) and a local iterative method (RNGDet++), the retrieval time of GLD-Road exhibited reductions of 40% and 92%, respectively, suggesting that GLD-Road achieves a pronounced improvement in road network extraction efficiency compared to existing methods. The experimental results are available at <ce:inter-ref xlink:href=\\\"https://github.com/ucas-dlg/GLD-Road\\\" xlink:type=\\\"simple\\\">https://github.com/ucas-dlg/GLD-Road</ce:inter-ref>.\",\"PeriodicalId\":50269,\"journal\":{\"name\":\"ISPRS Journal of Photogrammetry and Remote Sensing\",\"volume\":\"31 1\",\"pages\":\"\"},\"PeriodicalIF\":12.2000,\"publicationDate\":\"2025-08-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ISPRS Journal of Photogrammetry and Remote Sensing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://doi.org/10.1016/j.isprsjprs.2025.07.026\",\"RegionNum\":1,\"RegionCategory\":\"地球科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"GEOGRAPHY, PHYSICAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1016/j.isprsjprs.2025.07.026","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
GLD-Road: A global–local decoding road network extraction model for remote sensing images
Road networks are essential information for map updates, autonomous driving, and disaster response. However, manual annotation of road networks from remote sensing imagery is time-consuming and costly, whereas deep learning methods have gained attention for their efficiency and precision in road extraction. Current deep learning approaches for road network extraction fall into three main categories: postprocessing methods based on semantic segmentation results, global parallel methods and local iterative methods. Postprocessing methods introduce quantization errors, leading to higher overall road network inaccuracies; global parallel methods achieve high extraction efficiency but risk road node omissions; local iterative methods excel in node detection but have relatively lower extraction efficiency. To address the above limitations, We propose a two-stage road extraction model with global–local decoding, named GLD-Road, which possesses the high efficiency of global parallel methods and the strong node perception capability of local iterative methods, enabling a significant reduction in inference time while maintaining high-precision road network extraction. In the first stage, GLD-Road extracts the coordinates and direction descriptors of road nodes using global information from the entire input image. Subsequently, it connects adjacent nodes using a self-designed graph network module (Connect Module) to form the initial road network. In the second stage, based on the road endpoints contained in the initial road network, GLD-Road iteratively searches local images and the local grid map of the primary network to repair broken roads, ultimately producing a complete road network. Since the second stage only requires limited supplementary detection of locally missing nodes, GLD-Road significantly reduces the global iterative search range over the entire image, leading to a substantial reduction in retrieval time compared to local iterative methods. Finally, experimental results revealed that GLD-Road outperformed current state-of-the-art methods, achieving improvements of 1.9% and 0.67% in average path length similarity (APLS) on the City-Scale and SpaceNet3 datasets, respectively. Moreover, compared with those of a global parallel method (Sat2Graph) and a local iterative method (RNGDet++), the retrieval time of GLD-Road exhibited reductions of 40% and 92%, respectively, suggesting that GLD-Road achieves a pronounced improvement in road network extraction efficiency compared to existing methods. The experimental results are available at https://github.com/ucas-dlg/GLD-Road.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.