HSI-MFormer: Integrating Mamba and Transformer Experts for Hyperspectral Image Classification

IF 8.6 1区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC
Yan He;Bing Tu;Bo Liu;Jun Li;Antonio Plaza
{"title":"HSI-MFormer: Integrating Mamba and Transformer Experts for Hyperspectral Image Classification","authors":"Yan He;Bing Tu;Bo Liu;Jun Li;Antonio Plaza","doi":"10.1109/TGRS.2025.3564167","DOIUrl":null,"url":null,"abstract":"Hyperspectral image (HSI) classification is fundamental to numerous remote sensing applications, enabling detailed analysis of material properties and environmental conditions. Recent Mamba built upon selective state space models (SSMs) (S6) have demonstrated exceptional advantages in long-range sequence modeling with linear computational efficiency, while Transformer based on self-attention mechanisms is particularly adept at capturing short-range dependencies. To leverage the complementary strengths of these models, this article introduces a novel hybrid Mamba-Transformer framework (HSI-MFormer), effectively exploring the multiscale properties of hyperspectral data for HSI classification. Initially, a multiscale token generation (MTG) module is developed, which converts the HSI cube into multiple spatial-spectral token groups across different scales. To adequately capture fine-grained multiscale spatial-spectral patterns, an inner-scale transformer expert (ITE) is designed, which incorporates grouped self-attention operations to perform short-range sequence modeling within token groups at each scale. Meanwhile, a cross-scale Mamba expert (CME) is introduced, which integrates a cross-scale serialization mechanism and bidirectional Mamba block for long-range sequence modeling, further exploring the interactions and complementarity between token groups across different scales. Several hybrid strategies for integrating the ITE and CME are investigated to maximize their complementarity, including parallel, interval, and serial structures. Extensive experiments demonstrate that the proposed HSI-MFormer significantly outperforms the state-of-the-art Transformer-based and Mamba-based HSI classification methods. The code is available at <uri>http://github.com/tubingnuist/HSI-MFormer</uri>.","PeriodicalId":13213,"journal":{"name":"IEEE Transactions on Geoscience and Remote Sensing","volume":"63 ","pages":"1-16"},"PeriodicalIF":8.6000,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Geoscience and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10976442/","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

Abstract

Hyperspectral image (HSI) classification is fundamental to numerous remote sensing applications, enabling detailed analysis of material properties and environmental conditions. Recent Mamba built upon selective state space models (SSMs) (S6) have demonstrated exceptional advantages in long-range sequence modeling with linear computational efficiency, while Transformer based on self-attention mechanisms is particularly adept at capturing short-range dependencies. To leverage the complementary strengths of these models, this article introduces a novel hybrid Mamba-Transformer framework (HSI-MFormer), effectively exploring the multiscale properties of hyperspectral data for HSI classification. Initially, a multiscale token generation (MTG) module is developed, which converts the HSI cube into multiple spatial-spectral token groups across different scales. To adequately capture fine-grained multiscale spatial-spectral patterns, an inner-scale transformer expert (ITE) is designed, which incorporates grouped self-attention operations to perform short-range sequence modeling within token groups at each scale. Meanwhile, a cross-scale Mamba expert (CME) is introduced, which integrates a cross-scale serialization mechanism and bidirectional Mamba block for long-range sequence modeling, further exploring the interactions and complementarity between token groups across different scales. Several hybrid strategies for integrating the ITE and CME are investigated to maximize their complementarity, including parallel, interval, and serial structures. Extensive experiments demonstrate that the proposed HSI-MFormer significantly outperforms the state-of-the-art Transformer-based and Mamba-based HSI classification methods. The code is available at http://github.com/tubingnuist/HSI-MFormer.
HSI-MFormer:整合曼巴和变压器专家的高光谱图像分类
高光谱图像(HSI)分类是许多遥感应用的基础,可以对材料特性和环境条件进行详细分析。最近建立在选择性状态空间模型(ssm) (S6)上的Mamba在具有线性计算效率的远程序列建模方面表现出了非凡的优势,而基于自关注机制的Transformer特别擅长捕获短程依赖关系。为了利用这些模型的互补优势,本文引入了一种新的混合Mamba-Transformer框架(HSI- mformer),有效地探索了用于HSI分类的高光谱数据的多尺度特性。首先,开发了一个多尺度令牌生成(MTG)模块,将HSI立方体转换成不同尺度的多个空间光谱令牌组。为了充分捕获细粒度的多尺度空间光谱模式,设计了一个内尺度变压器专家(ITE),它包含分组自关注操作,在每个尺度的令牌组内执行短程序列建模。同时,引入跨尺度曼巴专家(CME),将跨尺度序列化机制和双向曼巴块集成到远程序列建模中,进一步探索不同尺度令牌组之间的相互作用和互补性。为了最大限度地提高二者的互补性,研究了几种混合集成策略,包括并行结构、区间结构和串联结构。大量的实验表明,提出的HSI- mformer显著优于最先进的基于变压器和基于mamba的HSI分类方法。代码可在http://github.com/tubingnuist/HSI-MFormer上获得。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
IEEE Transactions on Geoscience and Remote Sensing
IEEE Transactions on Geoscience and Remote Sensing 工程技术-地球化学与地球物理
CiteScore
11.50
自引率
28.00%
发文量
1912
审稿时长
4.0 months
期刊介绍: IEEE Transactions on Geoscience and Remote Sensing (TGRS) is a monthly publication that focuses on the theory, concepts, and techniques of science and engineering as applied to sensing the land, oceans, atmosphere, and space; and the processing, interpretation, and dissemination of this information.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信