{"title":"基于非配对SAR和红外图像交叉模态投影不变特征的车辆目标检测","authors":"Zhe Geng, Chongqi Xu, Chen Xin, Xiang Yu, Daiyin Zhu","doi":"10.1049/ell2.70336","DOIUrl":null,"url":null,"abstract":"<p>Synthetic aperture radar (SAR) automatic target recognition (ATR) is remarkably challenging since the SAR image defies the foundation for human and computer vision, i.e., the Gestalt perceptual principles. We propose to address this problem by fusing the target features reflected in SAR and infrared (IR) images via a novel dual-channel context-guided feature-alignment network (CGFAN) that is capable of fusing the cross-modality projective-invariant features extracted from unpaired SAR and IR images. First, region of interest (ROI) matching between SAR and IR images is realized based on special landmarks exhibiting consistent cross-modality features. After that, generative models trained with historical SAR and IR images are used to synthesize SAR images based on the IR images collected in real time for the current mission. Since SAR imaging takes more time than IR imaging, by using these synthesized SAR images as auxiliary data, the spatial-coverage rate in a typical collaborative SAR/IR ATR mission carried out by drone swarms is effectively improved. The proposed CGFAN is tested against the proprietary monostatic-bistatic circular SAR and IR dataset constructed by the researchers at our institution, which consists of nine types of military vehicles. Experimental results show that the proposed CGFAN offers better ATR performance than the baseline networks.A novel dual-channel CGFAN that is capable of fusing the cross-modality projective-invariant features extracted from unpaired SAR and IR images is proposed. First, ROI matching between SAR and IR images are realized based on special landmarks exhibiting consistent cross-modality features. After that, generative models trained with historical SAR and IR images are used to synthesize SAR images based on the IR images collected in real time for the current mission.</p>","PeriodicalId":11556,"journal":{"name":"Electronics Letters","volume":"61 1","pages":""},"PeriodicalIF":0.7000,"publicationDate":"2025-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/ell2.70336","citationCount":"0","resultStr":"{\"title\":\"Vehicle Target Detection Based on Cross-Modality Projective-Invariant Features Extracted from Unpaired SAR and Infrared Images\",\"authors\":\"Zhe Geng, Chongqi Xu, Chen Xin, Xiang Yu, Daiyin Zhu\",\"doi\":\"10.1049/ell2.70336\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Synthetic aperture radar (SAR) automatic target recognition (ATR) is remarkably challenging since the SAR image defies the foundation for human and computer vision, i.e., the Gestalt perceptual principles. We propose to address this problem by fusing the target features reflected in SAR and infrared (IR) images via a novel dual-channel context-guided feature-alignment network (CGFAN) that is capable of fusing the cross-modality projective-invariant features extracted from unpaired SAR and IR images. First, region of interest (ROI) matching between SAR and IR images is realized based on special landmarks exhibiting consistent cross-modality features. After that, generative models trained with historical SAR and IR images are used to synthesize SAR images based on the IR images collected in real time for the current mission. Since SAR imaging takes more time than IR imaging, by using these synthesized SAR images as auxiliary data, the spatial-coverage rate in a typical collaborative SAR/IR ATR mission carried out by drone swarms is effectively improved. The proposed CGFAN is tested against the proprietary monostatic-bistatic circular SAR and IR dataset constructed by the researchers at our institution, which consists of nine types of military vehicles. Experimental results show that the proposed CGFAN offers better ATR performance than the baseline networks.A novel dual-channel CGFAN that is capable of fusing the cross-modality projective-invariant features extracted from unpaired SAR and IR images is proposed. First, ROI matching between SAR and IR images are realized based on special landmarks exhibiting consistent cross-modality features. After that, generative models trained with historical SAR and IR images are used to synthesize SAR images based on the IR images collected in real time for the current mission.</p>\",\"PeriodicalId\":11556,\"journal\":{\"name\":\"Electronics Letters\",\"volume\":\"61 1\",\"pages\":\"\"},\"PeriodicalIF\":0.7000,\"publicationDate\":\"2025-06-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1049/ell2.70336\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Electronics Letters\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1049/ell2.70336\",\"RegionNum\":4,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Electronics Letters","FirstCategoryId":"5","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1049/ell2.70336","RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Vehicle Target Detection Based on Cross-Modality Projective-Invariant Features Extracted from Unpaired SAR and Infrared Images
Synthetic aperture radar (SAR) automatic target recognition (ATR) is remarkably challenging since the SAR image defies the foundation for human and computer vision, i.e., the Gestalt perceptual principles. We propose to address this problem by fusing the target features reflected in SAR and infrared (IR) images via a novel dual-channel context-guided feature-alignment network (CGFAN) that is capable of fusing the cross-modality projective-invariant features extracted from unpaired SAR and IR images. First, region of interest (ROI) matching between SAR and IR images is realized based on special landmarks exhibiting consistent cross-modality features. After that, generative models trained with historical SAR and IR images are used to synthesize SAR images based on the IR images collected in real time for the current mission. Since SAR imaging takes more time than IR imaging, by using these synthesized SAR images as auxiliary data, the spatial-coverage rate in a typical collaborative SAR/IR ATR mission carried out by drone swarms is effectively improved. The proposed CGFAN is tested against the proprietary monostatic-bistatic circular SAR and IR dataset constructed by the researchers at our institution, which consists of nine types of military vehicles. Experimental results show that the proposed CGFAN offers better ATR performance than the baseline networks.A novel dual-channel CGFAN that is capable of fusing the cross-modality projective-invariant features extracted from unpaired SAR and IR images is proposed. First, ROI matching between SAR and IR images are realized based on special landmarks exhibiting consistent cross-modality features. After that, generative models trained with historical SAR and IR images are used to synthesize SAR images based on the IR images collected in real time for the current mission.
期刊介绍:
Electronics Letters is an internationally renowned peer-reviewed rapid-communication journal that publishes short original research papers every two weeks. Its broad and interdisciplinary scope covers the latest developments in all electronic engineering related fields including communication, biomedical, optical and device technologies. Electronics Letters also provides further insight into some of the latest developments through special features and interviews.
Scope
As a journal at the forefront of its field, Electronics Letters publishes papers covering all themes of electronic and electrical engineering. The major themes of the journal are listed below.
Antennas and Propagation
Biomedical and Bioinspired Technologies, Signal Processing and Applications
Control Engineering
Electromagnetism: Theory, Materials and Devices
Electronic Circuits and Systems
Image, Video and Vision Processing and Applications
Information, Computing and Communications
Instrumentation and Measurement
Microwave Technology
Optical Communications
Photonics and Opto-Electronics
Power Electronics, Energy and Sustainability
Radar, Sonar and Navigation
Semiconductor Technology
Signal Processing
MIMO