MDNet: A Multi-modal Dual Branch Road Extraction Network Using Infrared Information

Yuxuan Gu
{"title":"MDNet: A Multi-modal Dual Branch Road Extraction Network Using Infrared Information","authors":"Yuxuan Gu","doi":"10.1109/ICGMRS55602.2022.9849226","DOIUrl":null,"url":null,"abstract":"Ambient occlusion and confusing terrain are difficult problems in remote sensing road extraction. In order to improve the accuracy of road extraction, this paper introduces infrared images as the basis for road extraction and proposes a multi-modal road extraction model (MDNet) suitable for additional infrared channel remote sensing images. First, this method adds a unidirectional D-LinkNet branch that is exactly the same as the backbone based on the D-LinkNet semantic segmentation network to learn infrared images and construct multi-modal image data information. Second, the attention mechanism was introduced in the multi-modal image processing stage, and a weighted model based on the attention mechanism was proposed to improve the utilization of information about the road. Finally, a road detection dataset composed of RGB-IR four-channel remote sensing images was built, and an experiment is conducted compared with some relatively advanced deep learning methods. The experimental results show that compared with DLinkNet with RGB remote sensing image as input, the mean intersection-over-union (mIoU) of D-LinkNet with RGB-IR remote sensing image as input is improved by about 4.3%, while MDNet is improved by 2.6% compared with the RGB-IR version of D-LinkNet. It can be seen that the multi-modal road extraction model MDNet combined with infrared information proposed in this paper can achieve more accurate road extraction from remote sensing images.","PeriodicalId":129909,"journal":{"name":"2022 3rd International Conference on Geology, Mapping and Remote Sensing (ICGMRS)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-04-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 3rd International Conference on Geology, Mapping and Remote Sensing (ICGMRS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICGMRS55602.2022.9849226","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Ambient occlusion and confusing terrain are difficult problems in remote sensing road extraction. In order to improve the accuracy of road extraction, this paper introduces infrared images as the basis for road extraction and proposes a multi-modal road extraction model (MDNet) suitable for additional infrared channel remote sensing images. First, this method adds a unidirectional D-LinkNet branch that is exactly the same as the backbone based on the D-LinkNet semantic segmentation network to learn infrared images and construct multi-modal image data information. Second, the attention mechanism was introduced in the multi-modal image processing stage, and a weighted model based on the attention mechanism was proposed to improve the utilization of information about the road. Finally, a road detection dataset composed of RGB-IR four-channel remote sensing images was built, and an experiment is conducted compared with some relatively advanced deep learning methods. The experimental results show that compared with DLinkNet with RGB remote sensing image as input, the mean intersection-over-union (mIoU) of D-LinkNet with RGB-IR remote sensing image as input is improved by about 4.3%, while MDNet is improved by 2.6% compared with the RGB-IR version of D-LinkNet. It can be seen that the multi-modal road extraction model MDNet combined with infrared information proposed in this paper can achieve more accurate road extraction from remote sensing images.
基于红外信息的多模态双支路提取网络
环境遮挡和地形混淆是遥感道路提取中的难题。为了提高道路提取的精度,本文引入红外图像作为道路提取的基础,提出了一种适用于附加红外通道遥感图像的多模式道路提取模型(MDNet)。首先,该方法在D-LinkNet语义分割网络的基础上,增加与主干完全相同的单向D-LinkNet分支,学习红外图像,构建多模态图像数据信息;其次,在多模态图像处理阶段引入注意机制,提出基于注意机制的加权模型,提高道路信息的利用率;最后,构建了由RGB-IR四通道遥感图像组成的道路检测数据集,并与一些较为先进的深度学习方法进行了对比实验。实验结果表明,与RGB遥感图像作为输入的DLinkNet相比,RGB- ir遥感图像作为输入的D-LinkNet的平均相交过联合(mIoU)提高了约4.3%,而MDNet与RGB- ir版本的D-LinkNet相比提高了2.6%。可以看出,本文提出的结合红外信息的多模式道路提取模型MDNet可以实现更精确的遥感图像道路提取。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信