AM2CFN: Assimilation Modality Mapping Guided Crossmodal Fusion Network for HSI and LiDAR Data Joint Classification

Yinbiao Lu;Wenbo Yu;Xintong Wei;Jiahui Huang
{"title":"AM2CFN: Assimilation Modality Mapping Guided Crossmodal Fusion Network for HSI and LiDAR Data Joint Classification","authors":"Yinbiao Lu;Wenbo Yu;Xintong Wei;Jiahui Huang","doi":"10.1109/LGRS.2024.3514179","DOIUrl":null,"url":null,"abstract":"Combining their complementary properties, using hyperspectral image (HSI) and light detection and ranging (LiDAR) data improves classification performance. Nevertheless, the heterogeneous capturing instruments and distribution characteristics of these two remote sensing (RS) modalities always limit their application scopes in on-ground observation-related domains. This heterogeneity hinders capturing the crossmodal connection for discriminant information extraction and exchange. In this letter, we propose an assimilation modality mapping guided crossmodal fusion network (AM2CFN) for HSI and LiDAR data joint classification. Our motivation is to explore one RS assimilation modality (RSAM) by exploiting one latent crossmodal mapping strategy from HSI and LiDAR data simultaneously to remove the effect of modality heterogeneity and contribute to information exchange. AM2CFN constructs one level-wise assimilating encoder to simulate modality heterogeneity and enhance regional consistency. Modality intrinsic features are captured in this encoder to provide knowledge for modality assimilation. Furthermore, one RSAM balancing HS and LiDAR properties is explored. AM2CFN constructs one RSAM reconstruction decoder for modality reconstruction and classification. Dual constraints based on solid angle and Kullback-Leibler divergence are considered to restrain the information exchange process toward the optimal direction. Experiments show that AM2CFN outperforms several state-of-the-art techniques qualitatively and quantitatively. AM2CFN increases the overall accuracy (OA) by 2.46% and 1.62% on average on the Houston and MUUFL datasets. The codes will be available at \n<uri>https://github.com/GEOywb/AM2CFN</uri>","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":0.0000,"publicationDate":"2024-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10787041/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Combining their complementary properties, using hyperspectral image (HSI) and light detection and ranging (LiDAR) data improves classification performance. Nevertheless, the heterogeneous capturing instruments and distribution characteristics of these two remote sensing (RS) modalities always limit their application scopes in on-ground observation-related domains. This heterogeneity hinders capturing the crossmodal connection for discriminant information extraction and exchange. In this letter, we propose an assimilation modality mapping guided crossmodal fusion network (AM2CFN) for HSI and LiDAR data joint classification. Our motivation is to explore one RS assimilation modality (RSAM) by exploiting one latent crossmodal mapping strategy from HSI and LiDAR data simultaneously to remove the effect of modality heterogeneity and contribute to information exchange. AM2CFN constructs one level-wise assimilating encoder to simulate modality heterogeneity and enhance regional consistency. Modality intrinsic features are captured in this encoder to provide knowledge for modality assimilation. Furthermore, one RSAM balancing HS and LiDAR properties is explored. AM2CFN constructs one RSAM reconstruction decoder for modality reconstruction and classification. Dual constraints based on solid angle and Kullback-Leibler divergence are considered to restrain the information exchange process toward the optimal direction. Experiments show that AM2CFN outperforms several state-of-the-art techniques qualitatively and quantitatively. AM2CFN increases the overall accuracy (OA) by 2.46% and 1.62% on average on the Houston and MUUFL datasets. The codes will be available at https://github.com/GEOywb/AM2CFN
基于同化模态映射的HSI与LiDAR数据联合分类跨模态融合网络
结合它们的互补特性,使用高光谱图像(HSI)和光探测和测距(LiDAR)数据可以提高分类性能。然而,这两种遥感方式的不同捕获手段和分布特点限制了它们在地面观测相关领域的应用范围。这种异质性阻碍了捕获跨模态连接以进行判别信息提取和交换。在这篇文章中,我们提出了一个同化模态映射引导的跨模态融合网络(AM2CFN),用于HSI和LiDAR数据的联合分类。我们的动机是通过同时利用HSI和LiDAR数据的一种潜在跨模态映射策略来探索一种RS同化模态(RSAM),以消除模态异质性的影响并促进信息交换。AM2CFN构建了一个逐级同化编码器,模拟模态异质性,增强区域一致性。该编码器捕获了情态固有特征,为情态同化提供了知识。此外,还探索了一种平衡HS和LiDAR特性的RSAM。AM2CFN构建了一个RSAM重构解码器,用于模态重构和分类。考虑了基于立体角和Kullback-Leibler散度的对偶约束,将信息交换过程约束在最优方向上。实验表明,AM2CFN在定性和定量上都优于几种最先进的技术。AM2CFN在休斯顿和MUUFL数据集上的总体精度(OA)平均提高2.46%和1.62%。这些代码可在https://github.com/GEOywb/AM2CFN上获得
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信