{"title":"基于多模态视图对比学习方法的小样本SAR目标识别","authors":"Yilin Li;Chengyu Wan;Xiaoyan Zhou;Tao Tang","doi":"10.1109/LGRS.2025.3557534","DOIUrl":null,"url":null,"abstract":"Self-supervised contrastive learning methods offer a promising approach to the small-sample synthetic aperture radar (SAR) automatic target recognition (ATR) problem by autonomously acquiring valuable visual representation from unlabeled data. However, current self-supervised contrastive learning methods primarily generate supervisory signals through augmented views of the original images, thereby underusing the rich information inherent in SAR images. To overcome this limitation, we integrate SAR targets’ geometric and physical properties, as captured in SAR target segmentation semantic maps and attribute scattering center reconstruction maps into the contrastive learning stage. Moreover, we propose a novel multimodal views’ contrastive learning method which contains two stages. In the contrastive learning stage, we leverage a large amount of unlabeled data for both intramodal and cross-modal contrastive learning, thereby transferring discriminative information from these two views to the original image features to learn the feature representation. In the supervised training stage, the linear classifier is trained using a small number of labeled samples to partition the feature representation space and migrate to the downstream recognition task. The experimental results demonstrate that the proposed method achieves superior recognition performance in SAR small-sample ATR tasks and exhibits robust generalization capabilities, thereby providing additional discriminative information that augments target representation.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":0.0000,"publicationDate":"2025-04-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Small-Sample SAR Target Recognition Using a Multimodal Views Contrastive Learning Method\",\"authors\":\"Yilin Li;Chengyu Wan;Xiaoyan Zhou;Tao Tang\",\"doi\":\"10.1109/LGRS.2025.3557534\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Self-supervised contrastive learning methods offer a promising approach to the small-sample synthetic aperture radar (SAR) automatic target recognition (ATR) problem by autonomously acquiring valuable visual representation from unlabeled data. However, current self-supervised contrastive learning methods primarily generate supervisory signals through augmented views of the original images, thereby underusing the rich information inherent in SAR images. To overcome this limitation, we integrate SAR targets’ geometric and physical properties, as captured in SAR target segmentation semantic maps and attribute scattering center reconstruction maps into the contrastive learning stage. Moreover, we propose a novel multimodal views’ contrastive learning method which contains two stages. In the contrastive learning stage, we leverage a large amount of unlabeled data for both intramodal and cross-modal contrastive learning, thereby transferring discriminative information from these two views to the original image features to learn the feature representation. In the supervised training stage, the linear classifier is trained using a small number of labeled samples to partition the feature representation space and migrate to the downstream recognition task. The experimental results demonstrate that the proposed method achieves superior recognition performance in SAR small-sample ATR tasks and exhibits robust generalization capabilities, thereby providing additional discriminative information that augments target representation.\",\"PeriodicalId\":91017,\"journal\":{\"name\":\"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society\",\"volume\":\"22 \",\"pages\":\"1-5\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-04-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10948483/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10948483/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Small-Sample SAR Target Recognition Using a Multimodal Views Contrastive Learning Method
Self-supervised contrastive learning methods offer a promising approach to the small-sample synthetic aperture radar (SAR) automatic target recognition (ATR) problem by autonomously acquiring valuable visual representation from unlabeled data. However, current self-supervised contrastive learning methods primarily generate supervisory signals through augmented views of the original images, thereby underusing the rich information inherent in SAR images. To overcome this limitation, we integrate SAR targets’ geometric and physical properties, as captured in SAR target segmentation semantic maps and attribute scattering center reconstruction maps into the contrastive learning stage. Moreover, we propose a novel multimodal views’ contrastive learning method which contains two stages. In the contrastive learning stage, we leverage a large amount of unlabeled data for both intramodal and cross-modal contrastive learning, thereby transferring discriminative information from these two views to the original image features to learn the feature representation. In the supervised training stage, the linear classifier is trained using a small number of labeled samples to partition the feature representation space and migrate to the downstream recognition task. The experimental results demonstrate that the proposed method achieves superior recognition performance in SAR small-sample ATR tasks and exhibits robust generalization capabilities, thereby providing additional discriminative information that augments target representation.