Jiepan Li , Wei He , Zhuohong Li , Yujun Guo , Hongyan Zhang
{"title":"克服遥感影像检测建筑物变化的不确定性挑战","authors":"Jiepan Li , Wei He , Zhuohong Li , Yujun Guo , Hongyan Zhang","doi":"10.1016/j.isprsjprs.2024.11.017","DOIUrl":null,"url":null,"abstract":"<div><div>Detecting building changes with multi-temporal remote sensing (RS) imagery at a very high resolution can help us understand urbanization and human activities, making informed decisions in urban planning, resource allocation, and infrastructure development. However, existing methods for building change detection (BCD) generally overlook critical uncertainty phenomena presented in RS imagery. Specifically, these uncertainties arise from two main sources: First, current building change detection datasets are designed primarily to detect changes in buildings, while changes in other land-cover classes are often classified as an unchanged background. Because of the manual labeling process, background elements that resemble buildings, such as roads and bridges, are at significant risk of being misclassified as building changes, introducing aleatoric uncertainty at the data level. Second, changes in parts of buildings that affect appearance, texture, or style without altering their semantic meaning, known as pseudo-changes, along with the imbalance between changed and unchanged samples, together lead to epistemic uncertainty at the model level. To address these challenges, we present an Uncertainty-Aware BCD (UA-BCD) framework. In detail, we employ a Siamese pyramid vision transformer to extract multi-level features from bi-temporal images, which are then decoded via a general decoder to obtain a coarse change map with inherent uncertainty. Subsequently, we introduce the aleatoric uncertainty estimation module to estimate the aleatoric uncertainty and embed it into the feature space. Then, a knowledge-guided feature enhancement module is developed to leverage the knowledge encoded in the coarse map to enhance the multi-level features and generate a refined change map. Finally, we propose an epistemic uncertainty estimator that takes the bi-temporal images and the refined change map as input and outputs an estimate of epistemic uncertainty. This estimation is supervised by the entropy calculated from the refined map, ensuring that the UA-BCD framework can produce a change map with lower epistemic uncertainty. To comprehensively validate the efficacy of the UA-BCD framework, we adopt a dual-perspective verification approach. Extensive experiments on five public building change datasets demonstrate the significant advantages of the proposed method over current state-of-the-art methods. Additionally, an application in Dongxihu District, Wuhan, China, confirms the outstanding performance of the proposed method in large-scale BCD. The source code of the project is available at <span><span>https://github.com/Henryjiepanli/UA-BCD</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"220 ","pages":"Pages 1-17"},"PeriodicalIF":10.6000,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Overcoming the uncertainty challenges in detecting building changes from remote sensing images\",\"authors\":\"Jiepan Li , Wei He , Zhuohong Li , Yujun Guo , Hongyan Zhang\",\"doi\":\"10.1016/j.isprsjprs.2024.11.017\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Detecting building changes with multi-temporal remote sensing (RS) imagery at a very high resolution can help us understand urbanization and human activities, making informed decisions in urban planning, resource allocation, and infrastructure development. However, existing methods for building change detection (BCD) generally overlook critical uncertainty phenomena presented in RS imagery. Specifically, these uncertainties arise from two main sources: First, current building change detection datasets are designed primarily to detect changes in buildings, while changes in other land-cover classes are often classified as an unchanged background. Because of the manual labeling process, background elements that resemble buildings, such as roads and bridges, are at significant risk of being misclassified as building changes, introducing aleatoric uncertainty at the data level. Second, changes in parts of buildings that affect appearance, texture, or style without altering their semantic meaning, known as pseudo-changes, along with the imbalance between changed and unchanged samples, together lead to epistemic uncertainty at the model level. To address these challenges, we present an Uncertainty-Aware BCD (UA-BCD) framework. In detail, we employ a Siamese pyramid vision transformer to extract multi-level features from bi-temporal images, which are then decoded via a general decoder to obtain a coarse change map with inherent uncertainty. Subsequently, we introduce the aleatoric uncertainty estimation module to estimate the aleatoric uncertainty and embed it into the feature space. Then, a knowledge-guided feature enhancement module is developed to leverage the knowledge encoded in the coarse map to enhance the multi-level features and generate a refined change map. Finally, we propose an epistemic uncertainty estimator that takes the bi-temporal images and the refined change map as input and outputs an estimate of epistemic uncertainty. This estimation is supervised by the entropy calculated from the refined map, ensuring that the UA-BCD framework can produce a change map with lower epistemic uncertainty. To comprehensively validate the efficacy of the UA-BCD framework, we adopt a dual-perspective verification approach. Extensive experiments on five public building change datasets demonstrate the significant advantages of the proposed method over current state-of-the-art methods. Additionally, an application in Dongxihu District, Wuhan, China, confirms the outstanding performance of the proposed method in large-scale BCD. The source code of the project is available at <span><span>https://github.com/Henryjiepanli/UA-BCD</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":50269,\"journal\":{\"name\":\"ISPRS Journal of Photogrammetry and Remote Sensing\",\"volume\":\"220 \",\"pages\":\"Pages 1-17\"},\"PeriodicalIF\":10.6000,\"publicationDate\":\"2025-02-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ISPRS Journal of Photogrammetry and Remote Sensing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S092427162400426X\",\"RegionNum\":1,\"RegionCategory\":\"地球科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"GEOGRAPHY, PHYSICAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S092427162400426X","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
Overcoming the uncertainty challenges in detecting building changes from remote sensing images
Detecting building changes with multi-temporal remote sensing (RS) imagery at a very high resolution can help us understand urbanization and human activities, making informed decisions in urban planning, resource allocation, and infrastructure development. However, existing methods for building change detection (BCD) generally overlook critical uncertainty phenomena presented in RS imagery. Specifically, these uncertainties arise from two main sources: First, current building change detection datasets are designed primarily to detect changes in buildings, while changes in other land-cover classes are often classified as an unchanged background. Because of the manual labeling process, background elements that resemble buildings, such as roads and bridges, are at significant risk of being misclassified as building changes, introducing aleatoric uncertainty at the data level. Second, changes in parts of buildings that affect appearance, texture, or style without altering their semantic meaning, known as pseudo-changes, along with the imbalance between changed and unchanged samples, together lead to epistemic uncertainty at the model level. To address these challenges, we present an Uncertainty-Aware BCD (UA-BCD) framework. In detail, we employ a Siamese pyramid vision transformer to extract multi-level features from bi-temporal images, which are then decoded via a general decoder to obtain a coarse change map with inherent uncertainty. Subsequently, we introduce the aleatoric uncertainty estimation module to estimate the aleatoric uncertainty and embed it into the feature space. Then, a knowledge-guided feature enhancement module is developed to leverage the knowledge encoded in the coarse map to enhance the multi-level features and generate a refined change map. Finally, we propose an epistemic uncertainty estimator that takes the bi-temporal images and the refined change map as input and outputs an estimate of epistemic uncertainty. This estimation is supervised by the entropy calculated from the refined map, ensuring that the UA-BCD framework can produce a change map with lower epistemic uncertainty. To comprehensively validate the efficacy of the UA-BCD framework, we adopt a dual-perspective verification approach. Extensive experiments on five public building change datasets demonstrate the significant advantages of the proposed method over current state-of-the-art methods. Additionally, an application in Dongxihu District, Wuhan, China, confirms the outstanding performance of the proposed method in large-scale BCD. The source code of the project is available at https://github.com/Henryjiepanli/UA-BCD.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.