Smail Ait El Asri, Ismail Negabi, Samir El Adib, N. Raissouni
{"title":"利用UNet和迁移学习增强遥感影像的建筑物提取","authors":"Smail Ait El Asri, Ismail Negabi, Samir El Adib, N. Raissouni","doi":"10.1080/1206212X.2023.2219117","DOIUrl":null,"url":null,"abstract":"Performing accurate extraction of buildings from remote sensing (RS) images is a crucial process with widespread applications in urban planning, disaster management, and urban monitoring. However, this task remains challenging due to the diversity and complexity of building shapes, sizes, and textures, as well as variations in lighting and weather conditions. These difficulties motivate our research to propose an improved approach for building extraction using UNet and transfer learning to address these challenges. In this work, we tested seven different backbone architectures within the UNet encoder and found that combining UNet with ResNet101 or ResNet152 yielded the best results. Based on these findings, we combined the superior performance of these base models to create a novel architecture, which achieved significant improvements over previous methods. Specifically, our method achieved a 1.33% increase in Intersection over Union (IoU) compared to the baseline UNet model. Furthermore, it demonstrated a superior performance with a 1.21% increase in IoU over UNet-ResNet101 and a 1.60% increase in IoU over UNet-ResNet152. We evaluated our proposed approach on the INRIA Aerial Image dataset and demonstrated its superiority. Our research addresses a critical need for accurate building extraction from RS images and overcomes the challenges posed by diverse building characteristics.","PeriodicalId":39673,"journal":{"name":"International Journal of Computers and Applications","volume":"442 1","pages":"413 - 419"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Enhancing building extraction from remote sensing images through UNet and transfer learning\",\"authors\":\"Smail Ait El Asri, Ismail Negabi, Samir El Adib, N. Raissouni\",\"doi\":\"10.1080/1206212X.2023.2219117\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Performing accurate extraction of buildings from remote sensing (RS) images is a crucial process with widespread applications in urban planning, disaster management, and urban monitoring. However, this task remains challenging due to the diversity and complexity of building shapes, sizes, and textures, as well as variations in lighting and weather conditions. These difficulties motivate our research to propose an improved approach for building extraction using UNet and transfer learning to address these challenges. In this work, we tested seven different backbone architectures within the UNet encoder and found that combining UNet with ResNet101 or ResNet152 yielded the best results. Based on these findings, we combined the superior performance of these base models to create a novel architecture, which achieved significant improvements over previous methods. Specifically, our method achieved a 1.33% increase in Intersection over Union (IoU) compared to the baseline UNet model. Furthermore, it demonstrated a superior performance with a 1.21% increase in IoU over UNet-ResNet101 and a 1.60% increase in IoU over UNet-ResNet152. We evaluated our proposed approach on the INRIA Aerial Image dataset and demonstrated its superiority. Our research addresses a critical need for accurate building extraction from RS images and overcomes the challenges posed by diverse building characteristics.\",\"PeriodicalId\":39673,\"journal\":{\"name\":\"International Journal of Computers and Applications\",\"volume\":\"442 1\",\"pages\":\"413 - 419\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-05-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Journal of Computers and Applications\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1080/1206212X.2023.2219117\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"Computer Science\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Computers and Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1080/1206212X.2023.2219117","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"Computer Science","Score":null,"Total":0}
Enhancing building extraction from remote sensing images through UNet and transfer learning
Performing accurate extraction of buildings from remote sensing (RS) images is a crucial process with widespread applications in urban planning, disaster management, and urban monitoring. However, this task remains challenging due to the diversity and complexity of building shapes, sizes, and textures, as well as variations in lighting and weather conditions. These difficulties motivate our research to propose an improved approach for building extraction using UNet and transfer learning to address these challenges. In this work, we tested seven different backbone architectures within the UNet encoder and found that combining UNet with ResNet101 or ResNet152 yielded the best results. Based on these findings, we combined the superior performance of these base models to create a novel architecture, which achieved significant improvements over previous methods. Specifically, our method achieved a 1.33% increase in Intersection over Union (IoU) compared to the baseline UNet model. Furthermore, it demonstrated a superior performance with a 1.21% increase in IoU over UNet-ResNet101 and a 1.60% increase in IoU over UNet-ResNet152. We evaluated our proposed approach on the INRIA Aerial Image dataset and demonstrated its superiority. Our research addresses a critical need for accurate building extraction from RS images and overcomes the challenges posed by diverse building characteristics.
期刊介绍:
The International Journal of Computers and Applications (IJCA) is a unique platform for publishing novel ideas, research outcomes and fundamental advances in all aspects of Computer Science, Computer Engineering, and Computer Applications. This is a peer-reviewed international journal with a vision to provide the academic and industrial community a platform for presenting original research ideas and applications. IJCA welcomes four special types of papers in addition to the regular research papers within its scope: (a) Papers for which all results could be easily reproducible. For such papers, the authors will be asked to upload "instructions for reproduction'''', possibly with the source codes or stable URLs (from where the codes could be downloaded). (b) Papers with negative results. For such papers, the experimental setting and negative results must be presented in detail. Also, why the negative results are important for the research community must be explained clearly. The rationale behind this kind of paper is that this would help researchers choose the correct approaches to solve problems and avoid the (already worked out) failed approaches. (c) Detailed report, case study and literature review articles about innovative software / hardware, new technology, high impact computer applications and future development with sufficient background and subject coverage. (d) Special issue papers focussing on a particular theme with significant importance or papers selected from a relevant conference with sufficient improvement and new material to differentiate from the papers published in a conference proceedings.