Zheng Zhang;Tingfa Xu;Peng Lou;Peng Lv;Tiehong Tian;Jianan Li
{"title":"基于双峰自适应卷积的多光谱点云分割","authors":"Zheng Zhang;Tingfa Xu;Peng Lou;Peng Lv;Tiehong Tian;Jianan Li","doi":"10.1109/LGRS.2025.3565739","DOIUrl":null,"url":null,"abstract":"Multispectral point cloud segmentation, leveraging both spatial and spectral information to classify individual points, is crucial for applications such as remote sensing, autonomous driving, and urban planning. However, existing methods primarily focus on spatial information and merge it with spectral data without fully considering their differences, limiting the effective use of spectral information. In this letter, we introduce a novel approach, bimodal adaptive convolution (BiMAConv), which fully exploits information from different modalities, based on the divide-and-conquer philosophy. Specifically, BiMAConv leverages the spectral features provided by the spectral information divergence (SID) and the weight information provided by the modal-weight block (MW-Block) module. The SID highlights slight differences in spectral information, providing detailed differential feature information. The MW-Block module utilizes an attention mechanism to combine generated features with the original point cloud, thereby generating weights to maintain learning balance sharply. In addition, we reconstruct a large-scale urban point cloud dataset GRSS_DFC_2018_3D based on dataset GRSS_DFC_2018 to advance the field of multispectral remote sensing point cloud, with a greater number of categories, more precise annotations, and registered multispectral channels. BiMAConv is fundamentally plug-and-play and supports different shared-multilayer perceptron (MLP) methods with almost no architectural changes. Extensive experiments on GRSS_DFC_2018_3D and Toronto-3D benchmarks demonstrate that our method significantly boosts the performance of popular detectors.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":0.0000,"publicationDate":"2025-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"BiMAConv: Bimodal Adaptive Convolution for Multispectral Point Cloud Segmentation\",\"authors\":\"Zheng Zhang;Tingfa Xu;Peng Lou;Peng Lv;Tiehong Tian;Jianan Li\",\"doi\":\"10.1109/LGRS.2025.3565739\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Multispectral point cloud segmentation, leveraging both spatial and spectral information to classify individual points, is crucial for applications such as remote sensing, autonomous driving, and urban planning. However, existing methods primarily focus on spatial information and merge it with spectral data without fully considering their differences, limiting the effective use of spectral information. In this letter, we introduce a novel approach, bimodal adaptive convolution (BiMAConv), which fully exploits information from different modalities, based on the divide-and-conquer philosophy. Specifically, BiMAConv leverages the spectral features provided by the spectral information divergence (SID) and the weight information provided by the modal-weight block (MW-Block) module. The SID highlights slight differences in spectral information, providing detailed differential feature information. The MW-Block module utilizes an attention mechanism to combine generated features with the original point cloud, thereby generating weights to maintain learning balance sharply. In addition, we reconstruct a large-scale urban point cloud dataset GRSS_DFC_2018_3D based on dataset GRSS_DFC_2018 to advance the field of multispectral remote sensing point cloud, with a greater number of categories, more precise annotations, and registered multispectral channels. BiMAConv is fundamentally plug-and-play and supports different shared-multilayer perceptron (MLP) methods with almost no architectural changes. Extensive experiments on GRSS_DFC_2018_3D and Toronto-3D benchmarks demonstrate that our method significantly boosts the performance of popular detectors.\",\"PeriodicalId\":91017,\"journal\":{\"name\":\"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society\",\"volume\":\"22 \",\"pages\":\"1-5\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-04-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10980282/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10980282/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
BiMAConv: Bimodal Adaptive Convolution for Multispectral Point Cloud Segmentation
Multispectral point cloud segmentation, leveraging both spatial and spectral information to classify individual points, is crucial for applications such as remote sensing, autonomous driving, and urban planning. However, existing methods primarily focus on spatial information and merge it with spectral data without fully considering their differences, limiting the effective use of spectral information. In this letter, we introduce a novel approach, bimodal adaptive convolution (BiMAConv), which fully exploits information from different modalities, based on the divide-and-conquer philosophy. Specifically, BiMAConv leverages the spectral features provided by the spectral information divergence (SID) and the weight information provided by the modal-weight block (MW-Block) module. The SID highlights slight differences in spectral information, providing detailed differential feature information. The MW-Block module utilizes an attention mechanism to combine generated features with the original point cloud, thereby generating weights to maintain learning balance sharply. In addition, we reconstruct a large-scale urban point cloud dataset GRSS_DFC_2018_3D based on dataset GRSS_DFC_2018 to advance the field of multispectral remote sensing point cloud, with a greater number of categories, more precise annotations, and registered multispectral channels. BiMAConv is fundamentally plug-and-play and supports different shared-multilayer perceptron (MLP) methods with almost no architectural changes. Extensive experiments on GRSS_DFC_2018_3D and Toronto-3D benchmarks demonstrate that our method significantly boosts the performance of popular detectors.