Multi-scale fusion pixel and instance contrastive self-supervised learning for semantic segmentation of high-resolution Earth surface images

IF 2.8 3区 地球科学 Q2 ASTRONOMY & ASTROPHYSICS
Bin Liu, Bing Li, Shuofeng Li
{"title":"Multi-scale fusion pixel and instance contrastive self-supervised learning for semantic segmentation of high-resolution Earth surface images","authors":"Bin Liu,&nbsp;Bing Li,&nbsp;Shuofeng Li","doi":"10.1016/j.asr.2024.10.026","DOIUrl":null,"url":null,"abstract":"<div><div>In the field of remote sensing (RS) image semantic segmentation, existing supervised learning methods rely on a large amount of labeled data, which limits their application scope. To solve this problem, we propose a new self-supervised learning method, called multi-scale fusion pixel and instance contrastive learning network (MPINet). This method first uses focal frequency loss to optimize the learning of high-level semantic information, and then strengthens the spatial information in the shallow feature map through multi-scale fusion pixel contrast learning, thereby improving the model’s ability to mine detailed features. Experiments are conducted on the International Society for Photogrammetry and Remote Sensing (ISPRS) Potsdam, LoveDA and UAVid datasets. The results show that our method achieves 49.89%, 70.98% and 63.55% in mIoU, OA and mF1 indicators on the ISPRS Potsdam dataset, which is 1.48%, 1.71% and 0.95% higher than the best method. On the LoveDA dataset, our method achieves 38.05%, 52.89% and 54.10% in mIoU, OA and mF1 indicators, which is 0.6%, 0.91% and 1.04% higher than the best method. On the UAVid dataset, our method achieved mIoU, OA, and mF1 scores of 57.92%, 76.97%, and 73.28%, respectively, representing improvements of 1.06%, 1.46%, and 1.15% compared to the best method. Experimental results show that the proposed method outperforms the existing optimal self-supervised learning method and the ImageNet pre-training method.</div></div>","PeriodicalId":50850,"journal":{"name":"Advances in Space Research","volume":"75 2","pages":"Pages 1854-1870"},"PeriodicalIF":2.8000,"publicationDate":"2025-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advances in Space Research","FirstCategoryId":"89","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0273117724010536","RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ASTRONOMY & ASTROPHYSICS","Score":null,"Total":0}
引用次数: 0

Abstract

In the field of remote sensing (RS) image semantic segmentation, existing supervised learning methods rely on a large amount of labeled data, which limits their application scope. To solve this problem, we propose a new self-supervised learning method, called multi-scale fusion pixel and instance contrastive learning network (MPINet). This method first uses focal frequency loss to optimize the learning of high-level semantic information, and then strengthens the spatial information in the shallow feature map through multi-scale fusion pixel contrast learning, thereby improving the model’s ability to mine detailed features. Experiments are conducted on the International Society for Photogrammetry and Remote Sensing (ISPRS) Potsdam, LoveDA and UAVid datasets. The results show that our method achieves 49.89%, 70.98% and 63.55% in mIoU, OA and mF1 indicators on the ISPRS Potsdam dataset, which is 1.48%, 1.71% and 0.95% higher than the best method. On the LoveDA dataset, our method achieves 38.05%, 52.89% and 54.10% in mIoU, OA and mF1 indicators, which is 0.6%, 0.91% and 1.04% higher than the best method. On the UAVid dataset, our method achieved mIoU, OA, and mF1 scores of 57.92%, 76.97%, and 73.28%, respectively, representing improvements of 1.06%, 1.46%, and 1.15% compared to the best method. Experimental results show that the proposed method outperforms the existing optimal self-supervised learning method and the ImageNet pre-training method.
求助全文
约1分钟内获得全文 求助全文
来源期刊
Advances in Space Research
Advances in Space Research 地学天文-地球科学综合
CiteScore
5.20
自引率
11.50%
发文量
800
审稿时长
5.8 months
期刊介绍: The COSPAR publication Advances in Space Research (ASR) is an open journal covering all areas of space research including: space studies of the Earth''s surface, meteorology, climate, the Earth-Moon system, planets and small bodies of the solar system, upper atmospheres, ionospheres and magnetospheres of the Earth and planets including reference atmospheres, space plasmas in the solar system, astrophysics from space, materials sciences in space, fundamental physics in space, space debris, space weather, Earth observations of space phenomena, etc. NB: Please note that manuscripts related to life sciences as related to space are no more accepted for submission to Advances in Space Research. Such manuscripts should now be submitted to the new COSPAR Journal Life Sciences in Space Research (LSSR). All submissions are reviewed by two scientists in the field. COSPAR is an interdisciplinary scientific organization concerned with the progress of space research on an international scale. Operating under the rules of ICSU, COSPAR ignores political considerations and considers all questions solely from the scientific viewpoint.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信