{"title":"MDNet: A Multi-modal Dual Branch Road Extraction Network Using Infrared Information","authors":"Yuxuan Gu","doi":"10.1109/ICGMRS55602.2022.9849226","DOIUrl":null,"url":null,"abstract":"Ambient occlusion and confusing terrain are difficult problems in remote sensing road extraction. In order to improve the accuracy of road extraction, this paper introduces infrared images as the basis for road extraction and proposes a multi-modal road extraction model (MDNet) suitable for additional infrared channel remote sensing images. First, this method adds a unidirectional D-LinkNet branch that is exactly the same as the backbone based on the D-LinkNet semantic segmentation network to learn infrared images and construct multi-modal image data information. Second, the attention mechanism was introduced in the multi-modal image processing stage, and a weighted model based on the attention mechanism was proposed to improve the utilization of information about the road. Finally, a road detection dataset composed of RGB-IR four-channel remote sensing images was built, and an experiment is conducted compared with some relatively advanced deep learning methods. The experimental results show that compared with DLinkNet with RGB remote sensing image as input, the mean intersection-over-union (mIoU) of D-LinkNet with RGB-IR remote sensing image as input is improved by about 4.3%, while MDNet is improved by 2.6% compared with the RGB-IR version of D-LinkNet. It can be seen that the multi-modal road extraction model MDNet combined with infrared information proposed in this paper can achieve more accurate road extraction from remote sensing images.","PeriodicalId":129909,"journal":{"name":"2022 3rd International Conference on Geology, Mapping and Remote Sensing (ICGMRS)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-04-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 3rd International Conference on Geology, Mapping and Remote Sensing (ICGMRS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICGMRS55602.2022.9849226","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Ambient occlusion and confusing terrain are difficult problems in remote sensing road extraction. In order to improve the accuracy of road extraction, this paper introduces infrared images as the basis for road extraction and proposes a multi-modal road extraction model (MDNet) suitable for additional infrared channel remote sensing images. First, this method adds a unidirectional D-LinkNet branch that is exactly the same as the backbone based on the D-LinkNet semantic segmentation network to learn infrared images and construct multi-modal image data information. Second, the attention mechanism was introduced in the multi-modal image processing stage, and a weighted model based on the attention mechanism was proposed to improve the utilization of information about the road. Finally, a road detection dataset composed of RGB-IR four-channel remote sensing images was built, and an experiment is conducted compared with some relatively advanced deep learning methods. The experimental results show that compared with DLinkNet with RGB remote sensing image as input, the mean intersection-over-union (mIoU) of D-LinkNet with RGB-IR remote sensing image as input is improved by about 4.3%, while MDNet is improved by 2.6% compared with the RGB-IR version of D-LinkNet. It can be seen that the multi-modal road extraction model MDNet combined with infrared information proposed in this paper can achieve more accurate road extraction from remote sensing images.