{"title":"Contextual boundary-aware network for semantic segmentation of complex land transportation point cloud scenes","authors":"Yanming Chen , Jiakang Xia , Xincan Zou , Ziting Xiao , Xin Tang , Yufu Zang , Dong Chen , Yueqian Shen","doi":"10.1016/j.isprsjprs.2025.09.006","DOIUrl":null,"url":null,"abstract":"<div><div>Semantic segmentation of land transportation scenes is critical for infrastructure maintenance and the advancement of intelligent transportation systems. Unlike traditional large-scale scenes, land transportation environments present intricate structural dependencies among infrastructure elements and pronounced class imbalance. To address these challenges, we propose a Gaussian-enhanced positional encoding block that leverages the Gaussian function’s intrinsic smoothing and reweighting properties to project relative positional information into a higher-dimensional space. By fusing this enhanced representation with the original positional encoding, the model gains a more nuanced understanding of spatial dependencies among infrastructures, thereby improving its capacity for semantic segmentation in complex land transportation scenes. Furthermore, we introduce the Multi-Context Interaction Module (MCIM) into the backbone network, varying the number of MCIMs across different network levels to strengthen inter-layer context interactions and mitigate error accumulation. To mitigate class imbalance and excessive object adhesion within the scene, we incorporate a boundary-aware class-balanced (BCB) hybrid loss function. Comprehensive experiments on three distinct land transportation datasets validate the effectiveness of our approach, with comparative analyses against state-of-the-art methods demonstrating its consistent superiority. Specifically, our method attains the highest mIoU (91.8%) and OA (96.7%) on the high-speed rail dataset ExpressRail, the highest mIoU (73.3%) on the traditional railway dataset SNCF, and the highest mF1-score (87.4%) on the urban road dataset Pairs3D. Codes are uploaded at: <span><span>https://github.com/Kange7/CoBa</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"230 ","pages":"Pages 18-31"},"PeriodicalIF":12.2000,"publicationDate":"2025-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271625003582","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Semantic segmentation of land transportation scenes is critical for infrastructure maintenance and the advancement of intelligent transportation systems. Unlike traditional large-scale scenes, land transportation environments present intricate structural dependencies among infrastructure elements and pronounced class imbalance. To address these challenges, we propose a Gaussian-enhanced positional encoding block that leverages the Gaussian function’s intrinsic smoothing and reweighting properties to project relative positional information into a higher-dimensional space. By fusing this enhanced representation with the original positional encoding, the model gains a more nuanced understanding of spatial dependencies among infrastructures, thereby improving its capacity for semantic segmentation in complex land transportation scenes. Furthermore, we introduce the Multi-Context Interaction Module (MCIM) into the backbone network, varying the number of MCIMs across different network levels to strengthen inter-layer context interactions and mitigate error accumulation. To mitigate class imbalance and excessive object adhesion within the scene, we incorporate a boundary-aware class-balanced (BCB) hybrid loss function. Comprehensive experiments on three distinct land transportation datasets validate the effectiveness of our approach, with comparative analyses against state-of-the-art methods demonstrating its consistent superiority. Specifically, our method attains the highest mIoU (91.8%) and OA (96.7%) on the high-speed rail dataset ExpressRail, the highest mIoU (73.3%) on the traditional railway dataset SNCF, and the highest mF1-score (87.4%) on the urban road dataset Pairs3D. Codes are uploaded at: https://github.com/Kange7/CoBa.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.