ISPRS Journal of Photogrammetry and Remote Sensing最新文献

筛选
英文 中文
SDCluster: A clustering based self-supervised pre-training method for semantic segmentation of remote sensing images
IF 10.6 1区 地球科学
ISPRS Journal of Photogrammetry and Remote Sensing Pub Date : 2025-03-07 DOI: 10.1016/j.isprsjprs.2025.02.021
Hanwen Xu , Chenxiao Zhang , Peng Yue , Kaixuan Wang
{"title":"SDCluster: A clustering based self-supervised pre-training method for semantic segmentation of remote sensing images","authors":"Hanwen Xu ,&nbsp;Chenxiao Zhang ,&nbsp;Peng Yue ,&nbsp;Kaixuan Wang","doi":"10.1016/j.isprsjprs.2025.02.021","DOIUrl":"10.1016/j.isprsjprs.2025.02.021","url":null,"abstract":"<div><div>Reducing the reliance of remote sensing semantic segmentation models on labeled training data is essential for practical model deployment. Self-supervised pre-training methods, which learn representations from unlabeled data by designing pretext tasks, provide an approach to address this requirement. One inconvenience of the currently contrastive learning-based and masked image modeling-based self-supervised methods is the difficulty in evaluating the quality of the pre-trained model without fine-tuning for semantic segmentation task. Hence, this paper proposes a pixel-level clustering-based self-supervised learning method, named SDCluster, which allows for a qualitative evaluation of the pre-trained model through visualizing the clustering results. Specifically, SDCluster extends the self-distillation framework to the pixel-level by incorporating the clustering assignment module. Then, clustering constraint modules, including prototype constraint module and semantic consistency constraint module, are designed to eliminate ineffective cluster prototypes and preserve the semantic information of ground objects. Benefiting from the correlation between pixel-level clustering and per-pixel classification of semantic segmentation, experimental results indicate that SDCluster exhibits competitive fine-tuning accuracy and robust few-shot segmentation capabilities when compared to prevalent self-supervised methods. Large-scale pre-training experiment and practical application experiment also prove the generalization ability and extensibility of the proposed method. The code and the dataset for practical application experiment are available at <span><span>https://github.com/openrsgis/SDCluster</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"223 ","pages":"Pages 1-14"},"PeriodicalIF":10.6,"publicationDate":"2025-03-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143563151","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FengYun-3 meteorological satellites’ microwave radiation Imagers enhance land surface temperature measurements across the diurnal cycle
IF 10.6 1区 地球科学
ISPRS Journal of Photogrammetry and Remote Sensing Pub Date : 2025-03-06 DOI: 10.1016/j.isprsjprs.2025.02.018
Yuyang Xiong , Tianjie Zhao , Haishen Lü , Zhiqing Peng , Jingyao Zheng , Yu Bai , Panpan Yao , Peng Guo , Peilin Song , Zushuai Wei , Ronghan Xu , Shengli Wu , Lixin Dong , Lin Chen , Na Xu , Xiuqing Hu , Peng Zhang , Letu Husi , Jiancheng Shi
{"title":"FengYun-3 meteorological satellites’ microwave radiation Imagers enhance land surface temperature measurements across the diurnal cycle","authors":"Yuyang Xiong ,&nbsp;Tianjie Zhao ,&nbsp;Haishen Lü ,&nbsp;Zhiqing Peng ,&nbsp;Jingyao Zheng ,&nbsp;Yu Bai ,&nbsp;Panpan Yao ,&nbsp;Peng Guo ,&nbsp;Peilin Song ,&nbsp;Zushuai Wei ,&nbsp;Ronghan Xu ,&nbsp;Shengli Wu ,&nbsp;Lixin Dong ,&nbsp;Lin Chen ,&nbsp;Na Xu ,&nbsp;Xiuqing Hu ,&nbsp;Peng Zhang ,&nbsp;Letu Husi ,&nbsp;Jiancheng Shi","doi":"10.1016/j.isprsjprs.2025.02.018","DOIUrl":"10.1016/j.isprsjprs.2025.02.018","url":null,"abstract":"<div><div>Land Surface Temperature (LST) is a vital meteorological variable for assessing hydrological, ecological, and climatological dynamics, as well as energy exchanges at the land–atmosphere interface. Accurate and frequent LST measurement is essential for meteorological satellites. However, existing retrieval algorithms often fail to capture the nuances of diurnal temperature variations. This study utilizes the exceptional diurnal sampling capabilities of the Microwave Radiation Imagers (MWRI) on China’s FengYun-3 (FY-3) satellites to improve LST measurements throughout the day. The objective is to develop a global algorithm that can distinguish between frozen and thawed states of near-surface landscape. This algorithm integrates multi-channel brightness temperature data and an array of microwave indices to enhance accuracy across diverse land cover types. Validation against in-situ measurements, alongside the comparative analysis with ERA5 and MODIS LST products demonstrate the algorithm’s high robustness. Results reveal a correlation coefficient exceeding 0.87 between FY-3 MWRI-derived LST and 5-cm soil temperature, with a root mean squared error (RMSE) near 4 K, except at 14:00 for FY-3D. The theoretical uncertainty, estimated using triple collocation analysis of the three LST datasets from FY-3 MWRI, ERA5 and MODIS, is less than 4 K for the majority of the globe. Additionally, the FY-3 MWRI exhibits reduced diurnal variation in LST as compared to MODIS LST, the peak temperatures recorded by FY-3 MWRI occur with a certain time lag relative to MODIS, and the diurnal temperature range is generally narrower, showcasing its adeptness in delineating diurnal temperature cycles when deployed across the FY-3B/C/D satellite constellation.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"222 ","pages":"Pages 204-224"},"PeriodicalIF":10.6,"publicationDate":"2025-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143548806","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mitigation of tropospheric turbulent delays in InSAR time series by incorporating a stochastic process
IF 10.6 1区 地球科学
ISPRS Journal of Photogrammetry and Remote Sensing Pub Date : 2025-03-05 DOI: 10.1016/j.isprsjprs.2025.02.028
Hailu Chen , Yunzhong Shen , Lei Zhang , Hongyu Liang , Tengfei Feng , Xinyou Song
{"title":"Mitigation of tropospheric turbulent delays in InSAR time series by incorporating a stochastic process","authors":"Hailu Chen ,&nbsp;Yunzhong Shen ,&nbsp;Lei Zhang ,&nbsp;Hongyu Liang ,&nbsp;Tengfei Feng ,&nbsp;Xinyou Song","doi":"10.1016/j.isprsjprs.2025.02.028","DOIUrl":"10.1016/j.isprsjprs.2025.02.028","url":null,"abstract":"<div><div>Tropospheric delays present a significant challenge to accurately mapping the Earth’s surface movements using interferometric synthetic aperture radar (InSAR). These delays are typically divided into stratified and turbulent components. While efforts have been made to address the stratified component, effectively mitigating turbulence remains an ongoing challenge. In response, this study proposes a joint model that compasses both the deterministic components and stochastic elements to account for the phases raised by turbulent delays in full InSAR time series. In the joint model, the deformation phases are parameterized by time-domain polynomial, while the turbulent delays are treated as spatially correlated stochastic variables, defined by spatial variance–covariance functions. Least Squares Collocation (LSC) and Variance-Covariance Estimation (VCE) are employed to solve this joint model, enabling simultaneous estimation of modelled deformation and turbulent mixing from full InSAR time series. The rationale is rooted in the distinct temporal dependencies of deformation and turbulent delay. Its efficacy and versatility are demonstrated using simulated and Sentinel-1 data from Hong Kong International Airport (China) and the Southern Valley of California (USA). In simulations, the root mean square error (RMSE) of the differential delays decreased from 2.4 to 0.8 cm. In the Southern Valley, comparison with 70 GPS measurements showed a 73.7 % reduction in mean RMSE, from 1.9 to 0.5 cm. These results confirm the effectiveness of this approach in mitigating tropospheric turbulence delays in the time domain.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"222 ","pages":"Pages 186-203"},"PeriodicalIF":10.6,"publicationDate":"2025-03-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143548805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TACMT: Text-aware cross-modal transformer for visual grounding on high-resolution SAR images
IF 10.6 1区 地球科学
ISPRS Journal of Photogrammetry and Remote Sensing Pub Date : 2025-03-02 DOI: 10.1016/j.isprsjprs.2025.02.022
Tianyang Li , Chao Wang , Sirui Tian , Bo Zhang , Fan Wu , Yixian Tang , Hong Zhang
{"title":"TACMT: Text-aware cross-modal transformer for visual grounding on high-resolution SAR images","authors":"Tianyang Li ,&nbsp;Chao Wang ,&nbsp;Sirui Tian ,&nbsp;Bo Zhang ,&nbsp;Fan Wu ,&nbsp;Yixian Tang ,&nbsp;Hong Zhang","doi":"10.1016/j.isprsjprs.2025.02.022","DOIUrl":"10.1016/j.isprsjprs.2025.02.022","url":null,"abstract":"<div><div>This paper introduces a novel task of visual grounding for high-resolution synthetic aperture radar images (SARVG). SARVG aims to identify the referred object in images through natural language instructions. While object detection on SAR images has been extensively investigated, identifying objects based on natural language remains under-explored. Due to the unique satellite view and side-look geometry, substantial expertise is often required to interpret objects, making it challenging to generalize across different sensors. Therefore, we propose to construct a dataset and develop multimodal deep learning models for the SARVG task. Our contributions can be summarized as follows. Using power transmission tower detection as an example, we have built a new benchmark of SARVG based on images from different SAR sensors to fully promote SARVG research. Subsequently, a novel text-aware cross-modal Transformer (TACMT) is proposed which follows DETR’s architecture. We develop a cross-modal encoder to enhance the visual features associated with the textual descriptions. Next, a text-aware query selection module is devised to select relevant context features as the decoder query. To retrieve the object from various scenes, we further design a cross-scale fusion module to fuse features from different levels for accurate target localization. Finally, extensive experiments on our dataset and widely used public datasets have demonstrated the effectiveness of our proposed model. This work provides valuable insights for SAR image interpretation. The code and dataset are available at <span><span>https://github.com/CAESAR-Radi/TACMT</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"222 ","pages":"Pages 152-166"},"PeriodicalIF":10.6,"publicationDate":"2025-03-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143526819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Time-Series models for ground subsidence and heave over permafrost in InSAR Processing: A comprehensive assessment and new improvement InSAR 处理中冻土层地面沉降和隆起的时间序列模型:全面评估和新的改进
IF 10.6 1区 地球科学
ISPRS Journal of Photogrammetry and Remote Sensing Pub Date : 2025-03-02 DOI: 10.1016/j.isprsjprs.2025.02.019
Chengyan Fan , Cuicui Mu , Lin Liu , Tingjun Zhang , Shichao Jia , Shengdi Wang , Wen Sun , Zhuoyi Zhao
{"title":"Time-Series models for ground subsidence and heave over permafrost in InSAR Processing: A comprehensive assessment and new improvement","authors":"Chengyan Fan ,&nbsp;Cuicui Mu ,&nbsp;Lin Liu ,&nbsp;Tingjun Zhang ,&nbsp;Shichao Jia ,&nbsp;Shengdi Wang ,&nbsp;Wen Sun ,&nbsp;Zhuoyi Zhao","doi":"10.1016/j.isprsjprs.2025.02.019","DOIUrl":"10.1016/j.isprsjprs.2025.02.019","url":null,"abstract":"<div><div>InSAR is an effective tool for indirectly monitoring large-scale hydrological-thermal dynamics of the active layer and permafrost by detecting the surface deformation. However, the conventional time-series models of InSAR technology do not consider the distinctive and pronounced seasonal characteristics of deformation over permafrost. Although permafrost-tailored models have been developed, their performance relative to the conventional models has not been assessed. In this study, we modify sinusoidal function and Stefan-equation-based models (permafrost-tailored) to better characterize surface deformation over permafrost, and assess advantages and limitations of these models for three application scenarios: filling time-series gaps for Small Baseline Subset (SBAS) inversion, deriving velocity and amplitude of deformation and selecting reference points automatically. The HyP3 interferograms generated from Sentinel-1 are utilized to analyze the surface deformation of the permafrost region over the upper reaches of the Heihe River Basin from 2017 to 2023. The result shows that adding a semi-annual component to the sinusoidal function can better capture the characteristics of ground surface deformation in permafrost regions. The modified Stefan-equation-based model performs well in those application scenarios, but it is only recommended for complex scenarios that conventional mathematical models cannot handle or for detailed simulations at individual points due to sophisticated data preparation and high computational cost. Furthermore, we find reference points can introduce substantial uncertainties into the deformation velocity and amplitude measurements, in comparison to the uncertainties derived from interferograms alone. The analysis of deformation amplitude and inter-annual velocity reveals that an ice-rich permafrost region, exhibiting a seasonal amplitude of 50–130 mm, is experiencing rapid degradation characterized by a subsidence velocity ranging from −10 to −20 mm/yr. Our study gives a permafrost-tailored modification and quantitative assessment on the InSAR time-series models. It can also serve as a reference and promotion for the application of InSAR technology in future permafrost research. The dataset and code are available at <span><span>https://github.com/Fanchengyan/FanInSAR</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"222 ","pages":"Pages 167-185"},"PeriodicalIF":10.6,"publicationDate":"2025-03-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143526820","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bounding box versus point annotation: The impact on deep learning performance for animal detection in aerial images
IF 10.6 1区 地球科学
ISPRS Journal of Photogrammetry and Remote Sensing Pub Date : 2025-02-27 DOI: 10.1016/j.isprsjprs.2025.02.017
Zeyu Xu , Tiejun Wang , Andrew K. Skidmore , Richard Lamprey , Shadrack Ngene
{"title":"Bounding box versus point annotation: The impact on deep learning performance for animal detection in aerial images","authors":"Zeyu Xu ,&nbsp;Tiejun Wang ,&nbsp;Andrew K. Skidmore ,&nbsp;Richard Lamprey ,&nbsp;Shadrack Ngene","doi":"10.1016/j.isprsjprs.2025.02.017","DOIUrl":"10.1016/j.isprsjprs.2025.02.017","url":null,"abstract":"<div><div>Bounding box and point annotations are widely used in deep learning-based animal detection from remote sensing imagery, yet their impact on model performance and training efficiency remains insufficiently explored. This study systematically evaluates the influence of these two annotation methods using aerial survey datasets of African elephants and antelopes across three commonly employed deep learning networks: YOLO, CenterNet, and U-Net. In addition, we assess the effect of image spatial resolution and the training efficiency associated with each annotation method. Our findings indicate that when using YOLO, there is no statistically significant difference in model accuracy between bounding box and point annotations. However, for CenterNet and U-Net, bounding box annotations consistently yield significantly higher accuracy compared to point-based annotations, with these trends remaining consistent across different spatial resolution ranges. Furthermore, training efficiency varies depending on the network and annotation method. While YOLO exhibits similar convergence speeds for both annotation types, U-Net models trained with bounding box annotations converge significantly faster, followed by CenterNet, where bounding box-based models also show improved convergence. These findings demonstrate that the choice of annotation method should be guided by the specific deep learning architecture employed. While point-based annotations are more cost-effective, their lower training efficiency in U-Net and CenterNet suggests that bounding box annotations are preferable when maximizing both accuracy and computational efficiency. Therefore, when selecting annotation strategies for animal detection in remote sensing applications, researchers should carefully balance detection accuracy, annotation cost, and training efficiency to optimize performance for specific task requirements.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"222 ","pages":"Pages 99-111"},"PeriodicalIF":10.6,"publicationDate":"2025-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143508621","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing LiDAR point cloud generation with BRDF-based appearance modelling
IF 10.6 1区 地球科学
ISPRS Journal of Photogrammetry and Remote Sensing Pub Date : 2025-02-27 DOI: 10.1016/j.isprsjprs.2025.02.010
Alfonso López, Carlos J. Ogayar, Rafael J. Segura, Juan C. Casas-Rosa
{"title":"Enhancing LiDAR point cloud generation with BRDF-based appearance modelling","authors":"Alfonso López,&nbsp;Carlos J. Ogayar,&nbsp;Rafael J. Segura,&nbsp;Juan C. Casas-Rosa","doi":"10.1016/j.isprsjprs.2025.02.010","DOIUrl":"10.1016/j.isprsjprs.2025.02.010","url":null,"abstract":"<div><div>This work presents an approach to generating LiDAR point clouds with empirical intensity data on a massively parallel scale. Our primary aim is to complement existing real-world LiDAR datasets by simulating a wide spectrum of attributes, ensuring our generated data can be directly compared to real point clouds. However, our emphasis lies in intensity data, which conventionally has been generated using non-photorealistic shading functions. In contrast, we represent surfaces with Bidirectional Reflectance Distribution Functions (BRDF) obtained through goniophotometer measurements. We also incorporate refractivity indices derived from prior research. Beyond this, we simulate other attributes commonly found in LiDAR datasets, including RGB values, normal vectors, GPS timestamps, semantic labels, instance IDs, and return data. Our simulations extend beyond terrestrial scenarios; we encompass mobile and aerial scans as well. Our results demonstrate the efficiency of our solution compared to other state-of-the-art simulators, achieving an average decrease in simulation time of 85.62%. Notably, our approach introduces greater variability in the generated intensity data, accounting for material properties and variations caused by the incident and viewing vectors. The source code is available on GitHub (<span><span>https://github.com/AlfonsoLRz/LiDAR_BRDF</span><svg><path></path></svg></span>).</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"222 ","pages":"Pages 79-98"},"PeriodicalIF":10.6,"publicationDate":"2025-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143508352","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LuoJiaHOG: A hierarchy oriented geo-aware image caption dataset for remote sensing image–text retrieval
IF 10.6 1区 地球科学
ISPRS Journal of Photogrammetry and Remote Sensing Pub Date : 2025-02-27 DOI: 10.1016/j.isprsjprs.2025.02.009
Yuanxin Zhao , Mi Zhang , Bingnan Yang , Zhan Zhang , Jujia Kang , Jianya Gong
{"title":"LuoJiaHOG: A hierarchy oriented geo-aware image caption dataset for remote sensing image–text retrieval","authors":"Yuanxin Zhao ,&nbsp;Mi Zhang ,&nbsp;Bingnan Yang ,&nbsp;Zhan Zhang ,&nbsp;Jujia Kang ,&nbsp;Jianya Gong","doi":"10.1016/j.isprsjprs.2025.02.009","DOIUrl":"10.1016/j.isprsjprs.2025.02.009","url":null,"abstract":"<div><div>Image–text retrieval (ITR) is crucial for making informed decisions in various remote sensing (RS) applications, including urban development and disaster prevention. However, creating ITR datasets that combine vision and language modalities requires extensive geo-spatial sampling, diverse categories, and detailed descriptions. To address these needs, we introduce the LuojiaHOG dataset, which is geospatially aware, label-extension-friendly, and features comprehensive captions. LuojiaHOG incorporates hierarchical spatial sampling, an extensible classification system aligned with Open Geospatial Consortium (OGC) standards, and detailed caption generation. Additionally, we propose a CLIP-based Image Semantic Enhancement Network (CISEN) to enhance sophisticated ITR capabilities. CISEN comprises dual-path knowledge transfer and progressive cross-modal feature fusion. The former transfers multimodal knowledge from a large, pretrained CLIP-like model, while the latter enhances visual-to-text alignment and fine-grained cross-modal feature integration. Comprehensive statistics on LuojiaHOG demonstrate its richness in sampling diversity, label quantity, and description granularity. Evaluations of LuojiaHOG using various state-of-the-art ITR models–including ALBEF, ALIGN, CLIP, FILIP, Wukong, GeoRSCLIP, and CISEN-employ second- and third-level labels. Adapter-tuning shows that CISEN outperforms others, achieving the highest scores with WMAP@5 rates of 88.47% and 87.28% on third-level ITR tasks, respectively. Moreover, CISEN shows improvements of approximately 1.3% and 0.9% in WMAP@5 over its baseline. When tested on previous RS ITR benchmarks, CISEN achieves performance close to the state-of-the-art methods. Pretraining on LuojiaHOG can further enhance retrieval results. These findings underscore the advancements of CISEN in accurately retrieving relevant information across images and texts. LuojiaHOG and CISEN can serve as foundational resources for future research on RS image–text alignment, supporting a broad spectrum of vision-language applications. The retrieval demo and dataset are available at:<span><span>https://huggingface.co/spaces/aleo1/LuojiaHOG-demo</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"222 ","pages":"Pages 130-151"},"PeriodicalIF":10.6,"publicationDate":"2025-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143512381","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Photogrammetric system of non-central refractive camera based on two-view 3D reconstruction
IF 10.6 1区 地球科学
ISPRS Journal of Photogrammetry and Remote Sensing Pub Date : 2025-02-27 DOI: 10.1016/j.isprsjprs.2025.02.016
Zhen Wu , Mingshu Nan , Haidong Zhang , Junzhou Huo , Shangqi Chen , Guanyu Chen , Zhang Cheng
{"title":"Photogrammetric system of non-central refractive camera based on two-view 3D reconstruction","authors":"Zhen Wu ,&nbsp;Mingshu Nan ,&nbsp;Haidong Zhang ,&nbsp;Junzhou Huo ,&nbsp;Shangqi Chen ,&nbsp;Guanyu Chen ,&nbsp;Zhang Cheng","doi":"10.1016/j.isprsjprs.2025.02.016","DOIUrl":"10.1016/j.isprsjprs.2025.02.016","url":null,"abstract":"<div><div>Due to the harsh construction environment of tunnels, the visual system must be fitted with a sphere cover of a certain thickness. The visual system with an optical sphere cover invalidates conventional measurement methods. Therefore, this paper provides a comprehensive visual measurement method using spherical glass refraction. First, the spherical glass refraction imaging is modeled using a geometry-driven camera model. Second, a three-parameter calibration method for the optical center offset unit vector, incident optical path offset distance, and optical center offset distance was proposed to accurately characterize refractive distortion. Then, a dynamic interval (DI) based on angle and depth constraints is introduced, and a DI-SGM algorithm utilizing the semi-global stereo matching method is developed to solve the polar constraint failure problem under refraction. Finally, an improved binocular parallax method that uses refraction image pairs is proposed and demonstrated to be effective and stable under spherical refraction using effectiveness and comprehensive data experiments. The constructed DI has narrow characteristics. The imaging model presented in this paper has an average space reconstruction error of only 0.087 mm. The maximum measurement error for sphere center distance is only 0.157 mm, which is comparable in accuracy to the case with no refraction. The proposed method provides an effective approach for applying visual measurement methods under refractive effects, thereby improving the visual system’s reliability in tunnel environments.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"222 ","pages":"Pages 112-129"},"PeriodicalIF":10.6,"publicationDate":"2025-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143508350","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive Discrepancy Masked Distillation for remote sensing object detection
IF 10.6 1区 地球科学
ISPRS Journal of Photogrammetry and Remote Sensing Pub Date : 2025-02-26 DOI: 10.1016/j.isprsjprs.2025.02.006
Cong Li, Gong Cheng, Junwei Han
{"title":"Adaptive Discrepancy Masked Distillation for remote sensing object detection","authors":"Cong Li,&nbsp;Gong Cheng,&nbsp;Junwei Han","doi":"10.1016/j.isprsjprs.2025.02.006","DOIUrl":"10.1016/j.isprsjprs.2025.02.006","url":null,"abstract":"<div><div>Knowledge distillation (KD) has become a promising technique for obtaining a performant student detector in remote sensing images by inheriting the knowledge from a heavy teacher detector. Unfortunately, not every pixel contributes (even detrimental) equally to the final KD performance. To dispel this problem, the existing methods usually derived a distillation mask to stress the valuable regions during KD. In this paper, we put forth Adaptive Discrepancy Masked Distillation (ADMD), a novel KD framework to explicitly localize the beneficial pixels. Our approach stems from the observation that the feature discrepancy between the teacher and student is the essential reason for their performance gap. With this regard, we make use of the feature discrepancy to determine which location causes the student to lag behind the teacher and then regulate the student to assign higher learning priority to them. Furthermore, we empirically observe that the discrepancy masked distillation leads to loss vanishing in later KD stages. To combat this issue, we introduce a simple yet practical weight-increasing module, in which the magnitude of KD loss is adaptively adjusted to ensure KD steadily contributes to student optimization. Comprehensive experiments on DIOR and DOTA across various dense detectors show that our ADMD consistently harvests remarkable performance gains, particularly under a prolonged distillation schedule, and exhibits superiority over state-of-the-art counterparts. Code and trained checkpoints will be made available at <span><span>https://github.com/swift1988</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"222 ","pages":"Pages 54-63"},"PeriodicalIF":10.6,"publicationDate":"2025-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143487188","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信