Fan Xu, Zhigao Shang, Qi-hui Wu, Xiaofei Zhang, Zebin Lin, Shuning Shao
{"title":"MUFNet: Toward Semantic Segmentation of Multi-spectral Remote Sensing Images","authors":"Fan Xu, Zhigao Shang, Qi-hui Wu, Xiaofei Zhang, Zebin Lin, Shuning Shao","doi":"10.1145/3508259.3508265","DOIUrl":null,"url":null,"abstract":"In this paper, a new convolutional neural network called multi-U fusion networks (MUFNet) is proposed for accurate semantic segmentation of multi-spectral remote sensing. Essentially, MUFNet is inspired by UNet, MFNet and CAM and fully combines their advantages. First, MUFNet introduces the skip connections into a multi-encoder-to-mono-decoder architecture, thereby facilitating the fusion of multi-scale and multi-channel spectral information. Second, the shortcut module in the decoder is revised by concatenating multiple spectral features from different encoders and then feeding the concatenated data into a CAM unit. Thus, the multi-spectral context semantics are fused and also the redundant feature maps are attention-compressed. Extensive simulations were conducted by testing UNet, UNet-4ch, MFNet and MUFNet on the 8400 RGB-NIR multi-spectral images with five categories from the GID image dataset. The visual results clearly showed that the proposed MUFNet can achieve more smoothing and complete segmentation performance than the other networks. Moreover, the measure values of mIoU, FWIoU and PA indicate that the proposed MUFNet can outperform the other networks in average semantic segmentation accuracy.","PeriodicalId":259099,"journal":{"name":"Proceedings of the 2021 4th Artificial Intelligence and Cloud Computing Conference","volume":"18 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2021 4th Artificial Intelligence and Cloud Computing Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3508259.3508265","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In this paper, a new convolutional neural network called multi-U fusion networks (MUFNet) is proposed for accurate semantic segmentation of multi-spectral remote sensing. Essentially, MUFNet is inspired by UNet, MFNet and CAM and fully combines their advantages. First, MUFNet introduces the skip connections into a multi-encoder-to-mono-decoder architecture, thereby facilitating the fusion of multi-scale and multi-channel spectral information. Second, the shortcut module in the decoder is revised by concatenating multiple spectral features from different encoders and then feeding the concatenated data into a CAM unit. Thus, the multi-spectral context semantics are fused and also the redundant feature maps are attention-compressed. Extensive simulations were conducted by testing UNet, UNet-4ch, MFNet and MUFNet on the 8400 RGB-NIR multi-spectral images with five categories from the GID image dataset. The visual results clearly showed that the proposed MUFNet can achieve more smoothing and complete segmentation performance than the other networks. Moreover, the measure values of mIoU, FWIoU and PA indicate that the proposed MUFNet can outperform the other networks in average semantic segmentation accuracy.