{"title":"基于无监督学习的单视点SAR图像三维目标重建","authors":"Yanni Wang;Hecheng Jia;Shilei Fu;Feng Xu","doi":"10.1109/TGRS.2025.3550633","DOIUrl":null,"url":null,"abstract":"Three-dimensional shape retrieval from synthetic aperture radar (SAR) images has long presented a significant research challenge, with single-view reconstruction proving even more complex due to constraints such as the scarcity of labeled data, limited sample diversity, and heightened sensitivity to radar scattering characteristics. Recently developed deep learning-based methods have made progress in single-view target reconstruction from SAR images. However, these methods still rely heavily on 3-D ground-truth supervision and fail to fully leverage SAR imaging mechanisms for 3-D reconstruction. To address these limitations, an end-to-end unsupervised single-view 3-D reconstruction framework based on a differentiable SAR renderer (DSR) is proposed, achieving precise reconstruction while eliminating the need for ground-truth data. Specifically, the encoder-decoder architecture effectively extracts 3-D and angular features, utilizing template deformation to preserve both fine details and global structures across scales, along with essential pose information for 3-D shape reconstruction. The reconstructed 3-D model is projected onto a 2-D plane, and pixel-level intersection over union (PIoU) loss is employed for unsupervised learning, enabling the extraction of discriminative latent structures and patterns. This approach effectively reduces low-frequency noise, sharpens critical edges, and enhances high-frequency details, improving spatial structure accuracy while minimizing shape distortions and height errors in complex targets. Extensive quantitative and qualitative experiments on both simulated and real datasets demonstrate the framework’s superior performance in single-view SAR 3-D target reconstruction, offering a promising solution with broad potential applications.","PeriodicalId":13213,"journal":{"name":"IEEE Transactions on Geoscience and Remote Sensing","volume":"63 ","pages":"1-12"},"PeriodicalIF":8.6000,"publicationDate":"2025-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Unsupervised Learning-Based 3-D Target Reconstruction From Single-View SAR Image\",\"authors\":\"Yanni Wang;Hecheng Jia;Shilei Fu;Feng Xu\",\"doi\":\"10.1109/TGRS.2025.3550633\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Three-dimensional shape retrieval from synthetic aperture radar (SAR) images has long presented a significant research challenge, with single-view reconstruction proving even more complex due to constraints such as the scarcity of labeled data, limited sample diversity, and heightened sensitivity to radar scattering characteristics. Recently developed deep learning-based methods have made progress in single-view target reconstruction from SAR images. However, these methods still rely heavily on 3-D ground-truth supervision and fail to fully leverage SAR imaging mechanisms for 3-D reconstruction. To address these limitations, an end-to-end unsupervised single-view 3-D reconstruction framework based on a differentiable SAR renderer (DSR) is proposed, achieving precise reconstruction while eliminating the need for ground-truth data. Specifically, the encoder-decoder architecture effectively extracts 3-D and angular features, utilizing template deformation to preserve both fine details and global structures across scales, along with essential pose information for 3-D shape reconstruction. The reconstructed 3-D model is projected onto a 2-D plane, and pixel-level intersection over union (PIoU) loss is employed for unsupervised learning, enabling the extraction of discriminative latent structures and patterns. This approach effectively reduces low-frequency noise, sharpens critical edges, and enhances high-frequency details, improving spatial structure accuracy while minimizing shape distortions and height errors in complex targets. Extensive quantitative and qualitative experiments on both simulated and real datasets demonstrate the framework’s superior performance in single-view SAR 3-D target reconstruction, offering a promising solution with broad potential applications.\",\"PeriodicalId\":13213,\"journal\":{\"name\":\"IEEE Transactions on Geoscience and Remote Sensing\",\"volume\":\"63 \",\"pages\":\"1-12\"},\"PeriodicalIF\":8.6000,\"publicationDate\":\"2025-03-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Geoscience and Remote Sensing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10924195/\",\"RegionNum\":1,\"RegionCategory\":\"地球科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Geoscience and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10924195/","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Unsupervised Learning-Based 3-D Target Reconstruction From Single-View SAR Image
Three-dimensional shape retrieval from synthetic aperture radar (SAR) images has long presented a significant research challenge, with single-view reconstruction proving even more complex due to constraints such as the scarcity of labeled data, limited sample diversity, and heightened sensitivity to radar scattering characteristics. Recently developed deep learning-based methods have made progress in single-view target reconstruction from SAR images. However, these methods still rely heavily on 3-D ground-truth supervision and fail to fully leverage SAR imaging mechanisms for 3-D reconstruction. To address these limitations, an end-to-end unsupervised single-view 3-D reconstruction framework based on a differentiable SAR renderer (DSR) is proposed, achieving precise reconstruction while eliminating the need for ground-truth data. Specifically, the encoder-decoder architecture effectively extracts 3-D and angular features, utilizing template deformation to preserve both fine details and global structures across scales, along with essential pose information for 3-D shape reconstruction. The reconstructed 3-D model is projected onto a 2-D plane, and pixel-level intersection over union (PIoU) loss is employed for unsupervised learning, enabling the extraction of discriminative latent structures and patterns. This approach effectively reduces low-frequency noise, sharpens critical edges, and enhances high-frequency details, improving spatial structure accuracy while minimizing shape distortions and height errors in complex targets. Extensive quantitative and qualitative experiments on both simulated and real datasets demonstrate the framework’s superior performance in single-view SAR 3-D target reconstruction, offering a promising solution with broad potential applications.
期刊介绍:
IEEE Transactions on Geoscience and Remote Sensing (TGRS) is a monthly publication that focuses on the theory, concepts, and techniques of science and engineering as applied to sensing the land, oceans, atmosphere, and space; and the processing, interpretation, and dissemination of this information.