{"title":"基于注意力的多级多尺度卷积网络PolSAR图像分类","authors":"Maryam Imani","doi":"10.1016/j.asr.2025.03.051","DOIUrl":null,"url":null,"abstract":"<div><div>Due to presence of heterogenous regions with materials and objects with different shapes and sizes in natural scenes, there are various contextual information in polarimetric synthetic aperture radar (PolSAR) images, which can be highlighted in the low-, medium-, or high-level features in different scales. To handle these challenges, the multi-scale and multi-level attention learning (MMAL) network is proposed for PolSAR image classification. A convolutional neural network (CNN) with six convolutional layers is introduced for hierarchical extraction of local contextual features in multiple levels. The cross-attention is used to find the relationships among low-level and high-level features and also among medium-level and high-level features. This process is repeated in multiple scales. Finally, the attention based multi-level and multi-scale features are fused to provide the classification map. An ablation study is done in several PolSAR images to show impact of different parts of the proposed network, which shows the superior efficiency of feature fusion in multiple levels and scales with taking to account the cross-attention among low–high and medium–high levels. The proposed MMAL network generally provides improved classification results compared to a CNN with the same structure and settings. For example, for the AIRSAR Flevoland image containing 15 class, with using 100 training samples per class, the overall accuracy of 96.15% with 2.74% increment with respect to the basic CNN is achieved where this improvement is statistically significant in term of the McNemars test. Moreover, the proposed method shows improvement compared to several state-of-the-art PolSAR classification methods.</div></div>","PeriodicalId":50850,"journal":{"name":"Advances in Space Research","volume":"75 11","pages":"Pages 7971-7986"},"PeriodicalIF":2.8000,"publicationDate":"2025-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Attention based multi-level and multi-scale convolutional network for PolSAR image classification\",\"authors\":\"Maryam Imani\",\"doi\":\"10.1016/j.asr.2025.03.051\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Due to presence of heterogenous regions with materials and objects with different shapes and sizes in natural scenes, there are various contextual information in polarimetric synthetic aperture radar (PolSAR) images, which can be highlighted in the low-, medium-, or high-level features in different scales. To handle these challenges, the multi-scale and multi-level attention learning (MMAL) network is proposed for PolSAR image classification. A convolutional neural network (CNN) with six convolutional layers is introduced for hierarchical extraction of local contextual features in multiple levels. The cross-attention is used to find the relationships among low-level and high-level features and also among medium-level and high-level features. This process is repeated in multiple scales. Finally, the attention based multi-level and multi-scale features are fused to provide the classification map. An ablation study is done in several PolSAR images to show impact of different parts of the proposed network, which shows the superior efficiency of feature fusion in multiple levels and scales with taking to account the cross-attention among low–high and medium–high levels. The proposed MMAL network generally provides improved classification results compared to a CNN with the same structure and settings. For example, for the AIRSAR Flevoland image containing 15 class, with using 100 training samples per class, the overall accuracy of 96.15% with 2.74% increment with respect to the basic CNN is achieved where this improvement is statistically significant in term of the McNemars test. Moreover, the proposed method shows improvement compared to several state-of-the-art PolSAR classification methods.</div></div>\",\"PeriodicalId\":50850,\"journal\":{\"name\":\"Advances in Space Research\",\"volume\":\"75 11\",\"pages\":\"Pages 7971-7986\"},\"PeriodicalIF\":2.8000,\"publicationDate\":\"2025-03-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Advances in Space Research\",\"FirstCategoryId\":\"89\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S027311772500290X\",\"RegionNum\":3,\"RegionCategory\":\"地球科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ASTRONOMY & ASTROPHYSICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advances in Space Research","FirstCategoryId":"89","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S027311772500290X","RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ASTRONOMY & ASTROPHYSICS","Score":null,"Total":0}
Attention based multi-level and multi-scale convolutional network for PolSAR image classification
Due to presence of heterogenous regions with materials and objects with different shapes and sizes in natural scenes, there are various contextual information in polarimetric synthetic aperture radar (PolSAR) images, which can be highlighted in the low-, medium-, or high-level features in different scales. To handle these challenges, the multi-scale and multi-level attention learning (MMAL) network is proposed for PolSAR image classification. A convolutional neural network (CNN) with six convolutional layers is introduced for hierarchical extraction of local contextual features in multiple levels. The cross-attention is used to find the relationships among low-level and high-level features and also among medium-level and high-level features. This process is repeated in multiple scales. Finally, the attention based multi-level and multi-scale features are fused to provide the classification map. An ablation study is done in several PolSAR images to show impact of different parts of the proposed network, which shows the superior efficiency of feature fusion in multiple levels and scales with taking to account the cross-attention among low–high and medium–high levels. The proposed MMAL network generally provides improved classification results compared to a CNN with the same structure and settings. For example, for the AIRSAR Flevoland image containing 15 class, with using 100 training samples per class, the overall accuracy of 96.15% with 2.74% increment with respect to the basic CNN is achieved where this improvement is statistically significant in term of the McNemars test. Moreover, the proposed method shows improvement compared to several state-of-the-art PolSAR classification methods.
期刊介绍:
The COSPAR publication Advances in Space Research (ASR) is an open journal covering all areas of space research including: space studies of the Earth''s surface, meteorology, climate, the Earth-Moon system, planets and small bodies of the solar system, upper atmospheres, ionospheres and magnetospheres of the Earth and planets including reference atmospheres, space plasmas in the solar system, astrophysics from space, materials sciences in space, fundamental physics in space, space debris, space weather, Earth observations of space phenomena, etc.
NB: Please note that manuscripts related to life sciences as related to space are no more accepted for submission to Advances in Space Research. Such manuscripts should now be submitted to the new COSPAR Journal Life Sciences in Space Research (LSSR).
All submissions are reviewed by two scientists in the field. COSPAR is an interdisciplinary scientific organization concerned with the progress of space research on an international scale. Operating under the rules of ICSU, COSPAR ignores political considerations and considers all questions solely from the scientific viewpoint.