Joel Amao-Oliva , Nils Foix-Colonier , Francescopaolo Sica
{"title":"Joint compression and despeckling by SAR representation learning","authors":"Joel Amao-Oliva , Nils Foix-Colonier , Francescopaolo Sica","doi":"10.1016/j.isprsjprs.2024.12.016","DOIUrl":null,"url":null,"abstract":"<div><div>Synthetic Aperture Radar (SAR) imagery is a powerful and widely used tool in a variety of remote sensing applications. The increasing number of SAR sensors makes it challenging to process and store such a large amount of data. In addition, as the flexibility and processing power of on-board electronics increases, the challenge of effectively transmitting large images to the ground becomes more tangible and pressing. In this paper, we present a method that uses self-supervised despeckling to learn a SAR image representation that is then used to perform image compression. The intuition that despeckling will additionally improve the compression task is based on the fact that the image representation used for despeckling forms an image prior that preserves the main image features while suppressing the spatially correlated noise component. The same learned image representation, which can already be seen as the output of a data reduction task, is further compressed in a lossless manner. While the two tasks can be solved separately, we propose to simultaneously training our model for despeckling and compression in a self-supervised and multi-objective fashion. The proposed network architecture avoids the use of skip connections by ensuring that the encoder and decoder share only the features generated at the lowest network level, namely the bridge, which is then further transformed into a bitstream. This differs from the usual network architectures used for despeckling, such as the commonly used Deep Residual U-Net. In this way, our network design allows compression and reconstruction to be performed at two different times and locations. The proposed method is trained and tested on real data from the TerraSAR-X sensor (downloaded from <span><span>https://earth.esa.int/eogateway/catalog/terrasar-x-esa-archive</span><svg><path></path></svg></span>). The experiments show that joint optimization can achieve performance beyond the state-of-the-art for both despeckling and compression, represented here by the <em>MERLIN</em> and <em>JPEG2000</em> algorithms, respectively. Furthermore, our method has been successfully tested against the cascade of these despeckling and compression algorithms, showing a better spatial and radiometric resolution, while achieving a better compression rate, e.g. a Peak Signal to Noise Ratio (PSNR) always higher than the comparison methods for any achieved bits-per-pixel (BPP) and specifically a PSNR gain of more than 2 dB by a compression rate of 0.7 BPP.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"220 ","pages":"Pages 524-534"},"PeriodicalIF":10.6000,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271624004866","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Synthetic Aperture Radar (SAR) imagery is a powerful and widely used tool in a variety of remote sensing applications. The increasing number of SAR sensors makes it challenging to process and store such a large amount of data. In addition, as the flexibility and processing power of on-board electronics increases, the challenge of effectively transmitting large images to the ground becomes more tangible and pressing. In this paper, we present a method that uses self-supervised despeckling to learn a SAR image representation that is then used to perform image compression. The intuition that despeckling will additionally improve the compression task is based on the fact that the image representation used for despeckling forms an image prior that preserves the main image features while suppressing the spatially correlated noise component. The same learned image representation, which can already be seen as the output of a data reduction task, is further compressed in a lossless manner. While the two tasks can be solved separately, we propose to simultaneously training our model for despeckling and compression in a self-supervised and multi-objective fashion. The proposed network architecture avoids the use of skip connections by ensuring that the encoder and decoder share only the features generated at the lowest network level, namely the bridge, which is then further transformed into a bitstream. This differs from the usual network architectures used for despeckling, such as the commonly used Deep Residual U-Net. In this way, our network design allows compression and reconstruction to be performed at two different times and locations. The proposed method is trained and tested on real data from the TerraSAR-X sensor (downloaded from https://earth.esa.int/eogateway/catalog/terrasar-x-esa-archive). The experiments show that joint optimization can achieve performance beyond the state-of-the-art for both despeckling and compression, represented here by the MERLIN and JPEG2000 algorithms, respectively. Furthermore, our method has been successfully tested against the cascade of these despeckling and compression algorithms, showing a better spatial and radiometric resolution, while achieving a better compression rate, e.g. a Peak Signal to Noise Ratio (PSNR) always higher than the comparison methods for any achieved bits-per-pixel (BPP) and specifically a PSNR gain of more than 2 dB by a compression rate of 0.7 BPP.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.