{"title":"Weather-aware autopilot: Domain generalization for point cloud semantic segmentation in diverse weather scenarios","authors":"","doi":"10.1016/j.isprsjprs.2024.09.006","DOIUrl":null,"url":null,"abstract":"<div><p>3D point cloud semantic segmentation, a pivotal task in fields such as autonomous driving and urban planning, confronts the challenge of performance degradation under adverse weather conditions. Current methodologies primarily focus on optimal weather scenarios, leaving a significant gap in handling various environmental adversities like fog, rain, and snow. To bridge this gap, we propose a comprehensive deep learning framework featuring unique components — an Adaptive Feature Normalization Module (AFNM) for effective normalization and calibration of features, a Dual-Attention Fusion Module (DAFM) for integrating cross-domain features, and a Proxy Label Generation Module (PLGM) for generating reliable proxy labels within the domain. Utilizing the SemanticKITTI and SynLiDAR datasets as source domains and the SemanticSTF dataset as the target domain, our model has been rigorously evaluated under varying weather conditions. When trained on the SemanticKITTI dataset as the source domain with the SemanticSTF dataset as the target, our approach surpasses the current state-of-the-art models by a margin of 6.2% in terms of overall mean Intersection over Union (mIoU) scores. Similarly, with the SynLiDAR dataset as the source and SemanticSTF as the target, our performance exceeds the best existing models by 3.4% in mIoU. These results substantiate the efficacy of our model in advancing the field of 3D semantic segmentation under diverse weather conditions, showcasing its notable robustness and superiority. The code is available at <span><span>https://github.com/J2DU/WADG-PointSeg</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":10.6000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271624003423","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
引用次数: 0
Abstract
3D point cloud semantic segmentation, a pivotal task in fields such as autonomous driving and urban planning, confronts the challenge of performance degradation under adverse weather conditions. Current methodologies primarily focus on optimal weather scenarios, leaving a significant gap in handling various environmental adversities like fog, rain, and snow. To bridge this gap, we propose a comprehensive deep learning framework featuring unique components — an Adaptive Feature Normalization Module (AFNM) for effective normalization and calibration of features, a Dual-Attention Fusion Module (DAFM) for integrating cross-domain features, and a Proxy Label Generation Module (PLGM) for generating reliable proxy labels within the domain. Utilizing the SemanticKITTI and SynLiDAR datasets as source domains and the SemanticSTF dataset as the target domain, our model has been rigorously evaluated under varying weather conditions. When trained on the SemanticKITTI dataset as the source domain with the SemanticSTF dataset as the target, our approach surpasses the current state-of-the-art models by a margin of 6.2% in terms of overall mean Intersection over Union (mIoU) scores. Similarly, with the SynLiDAR dataset as the source and SemanticSTF as the target, our performance exceeds the best existing models by 3.4% in mIoU. These results substantiate the efficacy of our model in advancing the field of 3D semantic segmentation under diverse weather conditions, showcasing its notable robustness and superiority. The code is available at https://github.com/J2DU/WADG-PointSeg.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.