Juncheng Han , Yuping Ye , Jixin Liang , Juan Zhao , Yi Chen , Xiujing Gao , Zhan Song
{"title":"High-precision underwater turbidity removal for 3D structured light reconstruction using division-of-focal-plane polarization","authors":"Juncheng Han , Yuping Ye , Jixin Liang , Juan Zhao , Yi Chen , Xiujing Gao , Zhan Song","doi":"10.1016/j.isprsjprs.2025.07.023","DOIUrl":"10.1016/j.isprsjprs.2025.07.023","url":null,"abstract":"<div><div>With the rapid advancement of marine technology, underwater 3D structured light reconstruction has emerged as a pivotal tool for ocean exploration, marine robotics, and underwater assembly. However, the aquatic environment typically contains numerous suspended particles that scatter and absorb signal light from the target, degrading underwater image quality and leading to unsatisfactory results with traditional image-based 3D reconstruction methods. To address this challenge, we have developed a temporal-multiplexing structured light system with polarization imaging to obtain dense and accurate target geometry information. Subsequently, we propose an underwater turbidity removal network based on this system. This network mitigates the impact of underwater turbidity on imaging accuracy and quality, significantly enhancing the precision and effectiveness of 3D reconstruction. Initially, we built a polarization system to capture fringe images of underwater objects at four different polarization angles. By calculating the second Stokes parameter and the degree of polarization, and then combining Canny features as input for the global contour capture block, we effectively extract the object’s contour information, restoring the fringe boundaries along its edges. Additionally, we introduce a fringe capture block to refine the boundaries of the fringe patterns, resulting in clearer internal reconstructions of the objects. By leveraging both the global contour capture block and the fringe capture block, we achieve high-precision underwater structured light reconstruction. Extensive experimental results demonstrate that our model achieves outstanding performance in terms of PSNR, SSIM, and LPIPS, with scores of 21.226, 0.969, and 0.050, respectively—significantly outperforming the reconstruction accuracy of state-of-the-art methods.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"228 ","pages":"Pages 453-466"},"PeriodicalIF":10.6,"publicationDate":"2025-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144711076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Kai Du , Yi Ma , Zhongwei Li , Zongchen Jiang , Rongjie Liu , Junfang Yang
{"title":"QOSM-U2CANet: a deep learning framework for normalized oil spill thickness and concentration mapping using multispectral satellite imagery","authors":"Kai Du , Yi Ma , Zhongwei Li , Zongchen Jiang , Rongjie Liu , Junfang Yang","doi":"10.1016/j.isprsjprs.2025.07.029","DOIUrl":"10.1016/j.isprsjprs.2025.07.029","url":null,"abstract":"<div><div>Marine oil spills pose significant ecological and socioeconomic threats, yet traditional remote sensing methods often fail to deliver accurate quantitative assessments of oil spill thickness and concentration across diverse environmental conditions. To address these limitations, this study presents QOSM-U2CANet, a novel deep learning framework that utilizes normalized thickness (<span><math><mrow><mi>nT</mi></mrow></math></span>) for non-emulsified oil spill (NEOS) and normalized concentration (<span><math><mrow><mi>nC</mi></mrow></math></span>) for emulsified oil spill (EOS) as physics-informed metrics derived from multispectral satellite imagery. The model incorporates a Residual Coordinate Attention U-Blocks (RCAU) to enhance spatial feature sensitivity and an L1 interval-constrained (L1-IC) loss function to improve detection precision in complex backgrounds. Evaluated across five oil spill-prone regions (Yellow Sea, South China Sea, Malacca Strait, Bay of Bengal, and Persian Gulf), QOSM-U2CANet achieves F1-scores of 0.855 for NEOS and 0.814 for EOS, with mean absolute errors (MAEs) of 0.077 for <span><math><mrow><mi>nT</mi></mrow></math></span> and 0.129 for <span><math><mrow><mi>nC</mi></mrow></math></span>, surpassing state-of-the-art models. The framework demonstrates resilience to environmental confounders and generalizes effectively across sensors with spatial resolutions of 10 to 50 m. Validation using historical Deepwater Horizon (DWH) oil spill retrospective estimates reveals a strong positive correlation (<span><math><mrow><msup><mrow><mi>R</mi></mrow><mn>2</mn></msup><mo>=</mo><mn>0.73</mn></mrow></math></span>) and an Intersection over Union (IoU) of 0.74 for oil spill distribution extent, confirming the model’s ability to accurately represent real-world oil spill spatial extents and volume distributions. By transitioning from categorical classification to continuous pixel-level mapping, QOSM-U2CANet provides detailed spatial distributions of oil thickness and concentration, enhancing emergency response strategies and ecological impact assessments. This advancement bridges qualitative detection and quantitative inversion, establishing a scalable data-driven foundation for global marine oil spill monitoring, with potential for further refinement under varying sun glint conditions and dynamic spill scenarios.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"228 ","pages":"Pages 420-437"},"PeriodicalIF":10.6,"publicationDate":"2025-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144711074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Chenchen Jiang , Huazhong Ren , Fengguang Li , Zhonghua Hong , Hongtao Huo , Junqiang Zhang , Jiuyuan Xin
{"title":"Object detection from aerial multi-angle thermal infrared remote sensing images: Dataset and method","authors":"Chenchen Jiang , Huazhong Ren , Fengguang Li , Zhonghua Hong , Hongtao Huo , Junqiang Zhang , Jiuyuan Xin","doi":"10.1016/j.isprsjprs.2025.07.024","DOIUrl":"10.1016/j.isprsjprs.2025.07.024","url":null,"abstract":"<div><div>Multi-angle thermal infrared (MATIR) remote sensing provides valuable day-night and multiple angular information that is of significant practical value in applications. However, multi-angle data heterogeneity is one of the core challenges in feature learning and scene understanding, which could severely degrade the model inference performance of deep neural networks. To address this issue, this study proposes a new fine-grained dataset and a unified method for the MATIR object detection task. In detail, the fine-grained MATIR object detection (MATIR-OD) dataset is captured by an unmanned aerial vehicle (UAV)-based platform, which offers significant advantages in terms of cost efficiency and exceptional maneuverability. The MATIR-OD dataset comprises 24 fine-grained and multi-angle data subsets, containing a total of 43,540 instances. Moreover, the unified MATIR object detection method, denoted as U-MATIR, includes the heterogeneous label space module and hybrid view cascade module. In the multi-angle object detection task, based on four public datasets and the proposed dataset, the all-angle experimental results show that the U-MATIR outperforms the ground- or aerial-view object detection models, increasing accuracy with an approximately 18–65% improvement in the mean Averaged Precision (mAP) metric, which exhibits notable robustness and generalization ability. In addition, the extensive experiments demonstrate the boundaries of robustness and generalization ability under 20–120 m and 30–90° fine-grained observation data. In particular, the optimal detection angle is defined as 60° under the above observation heights. The MATIR object detection dataset and unified method provide new insight for accurate multi-angle localization and achieve competitive detection performance.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"228 ","pages":"Pages 438-452"},"PeriodicalIF":10.6,"publicationDate":"2025-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144711075","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Simon Grieger , Martin Kappas , Susanne Karel , Philipp Koal , Tatjana Koukal , Markus Löw , Martin Zwanzig , Birgitta Putzenlechner
{"title":"Impact of forest disturbance derived from Sentinel-2 time series on Landsat 8/9 land surface temperature: The case of Norway spruce in Central Germany","authors":"Simon Grieger , Martin Kappas , Susanne Karel , Philipp Koal , Tatjana Koukal , Markus Löw , Martin Zwanzig , Birgitta Putzenlechner","doi":"10.1016/j.isprsjprs.2025.07.006","DOIUrl":"10.1016/j.isprsjprs.2025.07.006","url":null,"abstract":"<div><div>Forest cover and vitality loss is a global phenomenon. Areas of Norway spruce (<em>Picea abies</em> (L.) Karst.) in Central Germany were affected by widespread vitality and canopy cover loss in the years from 2018 due to drought stress and pest infestation. Such disturbances can favor higher land surface temperature (LST) on cloudless summer days. Regional assessment of LST in disturbed forest stands is challenging due to the spatial and temporal resolution of available products and various influences on the surface energy budget. To assess the effects of forest disturbance and topographic and pedological site factors on LST, a time series of the Landsat 8/9 Surface Temperature product was combined with a Sentinel-2-based forest disturbance monitoring framework. Results from three regions in Central Germany indicate a trend of elevated LST in disturbed areas of Norway spruce (median of LST differences of 4.4 K compared to undisturbed areas). Among topographic site factors, elevation exhibits the highest influence (median of LST differences between disturbed and undisturbed areas 1.2 K higher for highest areas compared to lowest). For pedological site factors, substrate shows the highest effect, modulating the median of LST differences by 2.9 K. Forest disturbance is accompanied by increased LST variance, possibly caused by different post-disturbance forest management practices. Air temperature at 15 cm shows highest agreement with LST and supports variation among management types. Identification of sites with a high risk of elevated LST is crucial for decision making in post-disturbance forest management, successful reforestation, and establishment of resilient forests.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"228 ","pages":"Pages 388-407"},"PeriodicalIF":10.6,"publicationDate":"2025-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144702143","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"BdFusion: Bi-directional visual-LiDAR fusion for resilient place recognition","authors":"Anbang Liang , Zhipeng Chen , Wen Xiong , Fanyi Meng , Yu Yin , Dejin Zhang , Qingquan Li","doi":"10.1016/j.isprsjprs.2025.07.022","DOIUrl":"10.1016/j.isprsjprs.2025.07.022","url":null,"abstract":"<div><div>Place recognition is an essential component for simultaneous localization and mapping (SLAM), as it is widely employed in loop closure detection to mitigate trajectory drifts. However, current works based on images or point clouds data are facing challenges in complex environments: single-modality methods may fail in degraded scenes; while conventional fusion methods simply combine multiple sensors data but ignore the problem that contributions of different features will dynamically change in different scenes. To improve the resilience of place recognition in complex environments, we propose a novel attention-based visual-LiDAR fusion method which is named BdFusion. In this work, a bi-directional attention module is proposed to improve the robustness of feature representation in changing environments, which performs explicit cross-modal feature interaction and enhancement by mining complementary features between 2D images and 3D point clouds. Furthermore, we design a feature fusion network that leverages multi-scale space and channel attention to comprehensively optimize the feature representation and fusion process, so as to learn the complementary advantages of multi modalities and perform adaptive feature fusion. Based on the fused feature, discriminative global descriptor is eventually constructed for place retrieval. We evaluate the proposed method on the self-built complex environment dataset and several public datasets. The experimental results show that our method outperforms existing state-of-the-art models such as PRFusion and AdaFusion on the challenging Szu dataset, achieving +1.6 Average Recall@1 (AR@1) and +0.6 Average Precision (AP), which effectively improves the accuracy and reliability of place recognition in complex environments. The code and dataset are publicly available at <span><span>https://github.com/ThomasLiangAB/BdFusion</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"228 ","pages":"Pages 408-419"},"PeriodicalIF":10.6,"publicationDate":"2025-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144702144","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"URSimulator: Human-perception-driven prompt tuning for enhanced virtual urban renewal via diffusion models","authors":"Chuanbo Hu , Shan Jia , Xin Li","doi":"10.1016/j.isprsjprs.2025.07.016","DOIUrl":"10.1016/j.isprsjprs.2025.07.016","url":null,"abstract":"<div><div>Tackling Urban Physical Disorder (UPD) – such as abandoned buildings, litter, messy vegetation, and graffiti – is essential, as it negatively impacts the safety, well-being, and psychological state of communities. Urban Renewal is the process of revitalizing these neglected and decayed areas within a city to improve their physical environment and quality of life for residents. Effective urban renewal efforts can transform these environments, enhancing their appeal and livability. However, current research lacks simulation tools that can quantitatively assess and visualize the impacts of urban renewal efforts, often relying on subjective judgments. Such simulation tools are essential for planning and implementing effective renewal strategies by providing a clear visualization of potential changes and their impacts. This paper presents a novel framework that addresses this gap by using human perception feedback to simulate the enhancement of street environment. We develop a prompt tuning approach that integrates text-driven Stable Diffusion with human perception feedback. This method iteratively edits local areas of street view images, aligning them more closely with human perceptions of beauty, liveliness, and safety. Our experiments show that this framework significantly improves people’s perceptions of urban environments, with increases of 17.60% in safety, 31.15% in beauty, and 28.82% in liveliness. In comparison, other advanced text-driven image editing methods like DiffEdit only achieve improvements of 2.31% in safety, 11.87% in beauty, and 15.84% in liveliness. We applied this framework across various virtual scenarios, including neighborhood improvement, building redevelopment, green space expansion, and community garden creation. The results demonstrate its effectiveness in simulating urban renewal, offering valuable insights for real-world urban planning and policy-making. This method not only enhances the visual appeal of neglected urban areas but also serves as a powerful tool for city planners and policymakers, ultimately improving urban landscapes and the quality of life for residents.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"228 ","pages":"Pages 356-369"},"PeriodicalIF":10.6,"publicationDate":"2025-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144695047","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ilia Parshakov , Derek Peddle , Karl Staenz , Jinkai Zhang , Craig Coburn , Howard Cheng
{"title":"UC–Change: a classification-based time series change detection technique for improved forest disturbance mapping using multi-sensor imagery","authors":"Ilia Parshakov , Derek Peddle , Karl Staenz , Jinkai Zhang , Craig Coburn , Howard Cheng","doi":"10.1016/j.isprsjprs.2025.07.028","DOIUrl":"10.1016/j.isprsjprs.2025.07.028","url":null,"abstract":"<div><div>Unsupervised Classification to Change (UC–Change) is a versatile technique that detects forest disturbances in satellite images by analyzing changes in the spatial distribution of spectral classes over time. This approach can fully utilize the spectral resolution of individual sensors without requiring atmospheric correction or radiometric normalization. Resulting multisensor capabilities set UC–Change apart from established time-series change detection methods, such as Continuous Change Detection and Classification (CCDC), LandTrendr, Composite2Change (C2C), and Global Forest Change (GFC). With the growing number of Earth observation satellites, the ability to utilize diverse datasets is increasingly important for extracting information relevant to sustainable natural resource management. The algorithm’s effectiveness is demonstrated using a dataset containing 275 Landsat and Sentinel–2 images acquired over a forested area in British Columbia, Canada, from 1972 to 2020. The 100 km × 100 km study site has been actively harvested in recent decades and experienced many wildfires and a mountain pine beetle (MPB) outbreak. The spatio-temporal accuracy of clearcut and fire-scar detection was assessed using the Vegetation Resources Inventory (VRI) and National Burned Area Composite (NBAC) products, respectively, and compared against the C2C 1985 – 2020, CCDC 2002 – 2019, and GFC 2001 – 2022 maps available online. Overall, the UC–Change algorithm detected 85.2 % of the reference VRI 1974 – 2018 cutblock pixels at a temporal resolution of ± 1 year (90.3 % at ± 3 years). It detected 86.0 % of 1985 – 2018 VRI pixels, outperforming C2C (58.8 %). For the period 2002 – 2018, UC–Change mapped 87.1 % of the reference cutblock pixels, exceeding C2C (54.5 %), CCDC (74.0 %), and GFC (70.4 %). UC–Change, C2C, CCDC, and GFC detected 71.0 %, 54.6 %, 37.4 %, and 67.2 % of 2006 – 2018 reference forest fire pixels, respectively. UC–Change provided improved forest harvest and fire-scar detection in areas heavily affected by the MPB outbreak and forests characterized by low canopy cover. It represents a new, fundamentally different approach to time-series analysis, suitable for independent use or in concert with existing methods.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"228 ","pages":"Pages 370-387"},"PeriodicalIF":10.6,"publicationDate":"2025-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144695048","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Biyuan Liu , Zhou Huang , Yanxi Li , Rongrong Gao , Huai-Xin Chen , Tian-Zhu Xiang
{"title":"HATFormer: Height-aware Transformer for multimodal 3D change detection","authors":"Biyuan Liu , Zhou Huang , Yanxi Li , Rongrong Gao , Huai-Xin Chen , Tian-Zhu Xiang","doi":"10.1016/j.isprsjprs.2025.06.022","DOIUrl":"10.1016/j.isprsjprs.2025.06.022","url":null,"abstract":"<div><div>Understanding the three-dimensional dynamics of the Earth’s surface is essential for urban planning and environmental monitoring. In the absence of consistent bitemporal 3D data, recent advancements in change detection have increasingly turned to combining multimodal data sources, including digital surface models (DSMs) and optical remote sensing imagery. However, significant inter-modal differences and intra-class variance — particularly with imbalances between foreground and background classes — continue to pose major challenges for achieving accurate change detection. To address these challenges, we propose a height-aware Transformer network, termed HATFormer, for multimodal semantic and height change detection, which explicitly correlates features across different modalities to reduce modality gaps and incorporates additional background supervision to mitigate foreground-to-background imbalances. Specifically, we first introduce a Background Height Estimation (BHE) module that incorporates height-awareness learning within the background to predict height information directly from lateral image features. This module enhances discriminative background feature learning and reduces the modality gap between monocular images and DSM data. To alleviate the interference of noisy background heights, a Height Uncertainty Suppression (HUS) module is designed to suppress the regions with height uncertainty. Secondly, we propose a Foreground Mask Estimation (FME) module to identify foreground change regions from DSM features, guided by discriminative background features. This module also acts as a regularizer, supporting more effective feature learning within the BHE module. Finally, an Auxiliary Feature Aggregation (AFA) module is designed to integrate features from the FME and BHE modules, which are then decoded by a multi-task decoder to generate precise change predictions. Extensive experiments on the Hi-BCD Plus and SMARS datasets demonstrate that our proposed method outperforms eight state-of-the-art methods, achieving superior performance in semantic and height change detection from multimodal bitemporal data. The code and dataset will be publicly available at: <span><span>https://github.com/HATFormer/HATFormer</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"228 ","pages":"Pages 340-355"},"PeriodicalIF":10.6,"publicationDate":"2025-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144686373","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Shuang Liu , Dong Li , Haibo Song , Caizhi Fan , Ke Li , Jun Wan , Ruining Liu
{"title":"SAR ship detection across different spaceborne platforms with confusion-corrected self-training and region-aware alignment framework","authors":"Shuang Liu , Dong Li , Haibo Song , Caizhi Fan , Ke Li , Jun Wan , Ruining Liu","doi":"10.1016/j.isprsjprs.2025.07.017","DOIUrl":"10.1016/j.isprsjprs.2025.07.017","url":null,"abstract":"<div><div>Synthetic Aperture Radar (SAR) ship detection is a vital technology for transforming reconnaissance data into actionable intelligence. As spaceborne SAR platforms increase, significant distribution shifts arise among SAR data from different platforms due to diverse imaging conditions and technical parameters. Traditional deep learning detectors, typically optimized for single-platform data, struggle with such shifts and annotation scarcity, limiting cross-platform applicability. Mainstream methods employ unsupervised domain adaptation (UDA) techniques to transfer detectors from a labeled source domain (existing platform data) to a novel unlabeled target domain (new platform data). However, the inherent complexity of SAR images, particularly strong background scattering regions, causes high confusion between ships and non-target regions, making these methods vulnerable to background interference and reducing their effectiveness in cross-platform detection. To alleviate this, we propose a <u>C</u>onfusion-Corrected <u>S</u>elf-Training with <u>R</u>egion-Aware <u>F</u>eature <u>A</u>lignment (CSRFA) framework for cross-platform SAR ship detection. First, a Confusion-corrected Self-training Mechanism (CSM) refines and corrects misclassified proposals to suppress background interference and enhance pseudo-label reliability on unlabeled target domains. Then, a Foreground Guidance Mechanism (FGM) further improves proposal quality by exploiting the consistency between region proposal classification and localization. Finally, a Region-Aware Feature Alignment (RAFA) module aligns ship regions based on RPN-generated foreground probabilities, enabling fine-grained, target-aware domain adaptation. Experiments on GF-3, SEN-1, and HRSID datasets show that CSRFA consistently outperforms existing UDA methods, achieving an average AP improvement of 2% across six cross-platform tasks compared to the second-best approach, demonstrating its robustness and adaptability for practical deployment.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"228 ","pages":"Pages 305-322"},"PeriodicalIF":10.6,"publicationDate":"2025-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144680105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Davide Lomeo , Stefan G.H. Simis , Nick Selmes , Anne D. Jungblut , Emma J. Tebbs
{"title":"Colour-informed ecoregion analysis highlights a satellite capability gap for spatially and temporally consistent freshwater cyanobacteria monitoring","authors":"Davide Lomeo , Stefan G.H. Simis , Nick Selmes , Anne D. Jungblut , Emma J. Tebbs","doi":"10.1016/j.isprsjprs.2025.07.030","DOIUrl":"10.1016/j.isprsjprs.2025.07.030","url":null,"abstract":"<div><div>Cyanobacteria blooms pose significant risks to water quality in freshwater ecosystems worldwide, with implications for human and animal health. Constructing consistent records of cyanobacteria dynamics in complex inland waters from satellite imagery remains challenged by discontinuous sensor capabilities, particularly with regard to spectral coverage. Comparing 11 satellite sensors, we show that the number and positioning of wavebands fundamentally alter bloom detection capability, with wavebands centred at 412, 620, 709, 754 and 779 nm proving most critical for capturing cyanobacteria dynamics. Specifically, analysis of observations from the Medium Resolution Imaging Spectrometer (MERIS) and Ocean and Land Colour Instrument (OLCI), coincident with the Moderate Resolution Imaging Spectroradiometer (MODIS) demonstrates how the spectral band configuration of the latter affects bloom detection. Using an Optical Water Types (OWT) library understood to capture cyanobacterial biomass through varying vertical mixing states, this analysis shows that MODIS can identify optically distinct conditions like surface accumulations but fails to resolve initial bloom evolution in well-mixed conditions, particularly in optically complex regions. Investigation of coherent ecoregions formed using Self-organising Maps trained on OWT membership scores confirm that MODIS captures broad spatial patterns seen with more capable sensors but compresses optical gradients into fewer optical types. These constraints have significant implications for interpreting spatial–temporal dynamics of cyanobacteria in large waterbodies, particularly during 2012–2016 when MERIS and OLCI sensors were absent, and small waterbodies, where high spatial resolution sensors not originally design to study water are used. In addition, these findings underscore the importance of key wavebands in future sensor design and the development of approaches to maintain consistent long-term records across evolving satellite capabilities. Our findings suggest that attempts at quantitatively harmonising cyanobacteria bloom detection across sensors may not be ecologically appropriate unless these observation biases are addressed. For example, analysing the frequency and intensity of surfacing blooms, while considering the meteorological factors that may drive these phenomena, could be considered over decadal timescales, whereas trend analysis of mixed-column biomass should only concern appropriate sensor observation periods.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"228 ","pages":"Pages 323-339"},"PeriodicalIF":10.6,"publicationDate":"2025-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144680018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}