Mamba-Wavelet Cross-Modal Fusion Network With Graph Pooling for Hyperspectral and LiDAR Data Joint Classification

Daxiang Li;Bingying Li;Ying Liu
{"title":"Mamba-Wavelet Cross-Modal Fusion Network With Graph Pooling for Hyperspectral and LiDAR Data Joint Classification","authors":"Daxiang Li;Bingying Li;Ying Liu","doi":"10.1109/LGRS.2025.3576778","DOIUrl":null,"url":null,"abstract":"Recently, with the rapid development of deep learning, the collaborative classification of hyperspectral image (HSI) and light detection and ranging (LiDAR) image has become a research hotspot in remote sensing (RS) technology. However, existing methods either only consider complementary learning of spatial-domain information or do not take into account the intrinsic dependencies between pixels and overlook the importance difference of pixels. In this letter, we propose a mamba-wavelet cross-modal fusion network with graph pooling (MW-CMFNet) for HSI and LiDAR joint classification. First, a two-branch feature extraction (TBFE) is used to extract spatial and spectral features. Then, in order to dig deeper into the complementary information of different modalities and fully fuse them under the guidance of frequency-domain information, a mamba-wavelet cross-modal feature fusion (MW-CMFF) module is devised, it aims to utilize mamba’s outstanding long-range modeling ability to learn complementary information in the spatial and frequency domains, Finally, the graph pooling module is designed to sense the intrinsic dependencies of neighboring pixels and explore the importance difference of pixels, rather than assigning the same weight to different pixels. Experiments on the Houston2013 and Trento datasets show that the MW-CMFNet achieves higher classification accuracy compared to other state-of-the-art methods.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":0.0000,"publicationDate":"2025-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/11026015/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Recently, with the rapid development of deep learning, the collaborative classification of hyperspectral image (HSI) and light detection and ranging (LiDAR) image has become a research hotspot in remote sensing (RS) technology. However, existing methods either only consider complementary learning of spatial-domain information or do not take into account the intrinsic dependencies between pixels and overlook the importance difference of pixels. In this letter, we propose a mamba-wavelet cross-modal fusion network with graph pooling (MW-CMFNet) for HSI and LiDAR joint classification. First, a two-branch feature extraction (TBFE) is used to extract spatial and spectral features. Then, in order to dig deeper into the complementary information of different modalities and fully fuse them under the guidance of frequency-domain information, a mamba-wavelet cross-modal feature fusion (MW-CMFF) module is devised, it aims to utilize mamba’s outstanding long-range modeling ability to learn complementary information in the spatial and frequency domains, Finally, the graph pooling module is designed to sense the intrinsic dependencies of neighboring pixels and explore the importance difference of pixels, rather than assigning the same weight to different pixels. Experiments on the Houston2013 and Trento datasets show that the MW-CMFNet achieves higher classification accuracy compared to other state-of-the-art methods.
基于图池的mamba -小波交叉模态融合网络用于高光谱和激光雷达数据联合分类
近年来,随着深度学习的快速发展,高光谱图像(HSI)与激光探测与测距(LiDAR)图像的协同分类已成为遥感技术的研究热点。然而,现有的方法要么只考虑了空域信息的互补学习,要么没有考虑像素之间的内在依赖关系,忽略了像素之间的重要性差异。在这篇文章中,我们提出了一种带有图池的曼巴-小波跨模融合网络(MW-CMFNet),用于HSI和LiDAR联合分类。首先,采用双分支特征提取(TBFE)方法提取空间和光谱特征;然后,为了更深入地挖掘不同模态的互补信息,并在频域信息的指导下进行充分融合,设计了曼巴-小波跨模态特征融合(MW-CMFF)模块,旨在利用曼巴出色的远程建模能力来学习空间域和频域的互补信息。图池化模块旨在感知相邻像素之间的内在依赖关系,并探索像素之间的重要性差异,而不是为不同的像素分配相同的权重。在Houston2013和Trento数据集上的实验表明,MW-CMFNet的分类精度高于其他先进方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信