{"title":"基于深度融合多尺度特征的高分辨率遥感影像道路提取并行双分支模型","authors":"Guoqing Zhou;Haiyang Zhi;Ertao Gao;Yanling Lu;Jianjun Chen;Yuhang Bai;Xiao Zhou","doi":"10.1109/JSTARS.2025.3555636","DOIUrl":null,"url":null,"abstract":"The existing encoder–decoder model, or encoder–decoder model with atrous convolutions, has exposed its limitations under diverse environments, such as road scales, shadows, building occlusions, and vegetation in high-resolution remote sensing images. Therefore, this article introduces a dual-branch deep fusion network, named “DeepU-Net,” for obtaining global and local information in parallel. Two novel modules are designed: 1) the spatial and coordinate squeeze-and-excitation fusion attention module that enhances the focus on spatial positions and target channel information; and 2) the efficient multiscale convolutional attention module that can boost the competence to tackle multiscale road information. The validation of the proposed model is conducted using two datasets, CHN6-CUG and DeepGlobe, which are from urban and rural areas, respectively. A comparative analysis with the six commonly used models, including U-Net, PSPNet, DeepLabv3+, HRNet, CoANet, and SegFormer, is conducted. The experimental results reveal that the introduced model achieves mean intersection over union scores of 83.18% and 81.43%, which are averagely improved by 1.93% and 1.02%, respectively, for the two datasets, when compared with the six commonly used models. The outcomes suggest that the introduced model achieves a greater accuracy than the six extensively applied models do.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"18 ","pages":"9448-9463"},"PeriodicalIF":4.7000,"publicationDate":"2025-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10945378","citationCount":"0","resultStr":"{\"title\":\"DeepU-Net: A Parallel Dual-Branch Model for Deeply Fusing Multiscale Features for Road Extraction From High-Resolution Remote Sensing Images\",\"authors\":\"Guoqing Zhou;Haiyang Zhi;Ertao Gao;Yanling Lu;Jianjun Chen;Yuhang Bai;Xiao Zhou\",\"doi\":\"10.1109/JSTARS.2025.3555636\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The existing encoder–decoder model, or encoder–decoder model with atrous convolutions, has exposed its limitations under diverse environments, such as road scales, shadows, building occlusions, and vegetation in high-resolution remote sensing images. Therefore, this article introduces a dual-branch deep fusion network, named “DeepU-Net,” for obtaining global and local information in parallel. Two novel modules are designed: 1) the spatial and coordinate squeeze-and-excitation fusion attention module that enhances the focus on spatial positions and target channel information; and 2) the efficient multiscale convolutional attention module that can boost the competence to tackle multiscale road information. The validation of the proposed model is conducted using two datasets, CHN6-CUG and DeepGlobe, which are from urban and rural areas, respectively. A comparative analysis with the six commonly used models, including U-Net, PSPNet, DeepLabv3+, HRNet, CoANet, and SegFormer, is conducted. The experimental results reveal that the introduced model achieves mean intersection over union scores of 83.18% and 81.43%, which are averagely improved by 1.93% and 1.02%, respectively, for the two datasets, when compared with the six commonly used models. The outcomes suggest that the introduced model achieves a greater accuracy than the six extensively applied models do.\",\"PeriodicalId\":13116,\"journal\":{\"name\":\"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing\",\"volume\":\"18 \",\"pages\":\"9448-9463\"},\"PeriodicalIF\":4.7000,\"publicationDate\":\"2025-03-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10945378\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10945378/\",\"RegionNum\":2,\"RegionCategory\":\"地球科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10945378/","RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
DeepU-Net: A Parallel Dual-Branch Model for Deeply Fusing Multiscale Features for Road Extraction From High-Resolution Remote Sensing Images
The existing encoder–decoder model, or encoder–decoder model with atrous convolutions, has exposed its limitations under diverse environments, such as road scales, shadows, building occlusions, and vegetation in high-resolution remote sensing images. Therefore, this article introduces a dual-branch deep fusion network, named “DeepU-Net,” for obtaining global and local information in parallel. Two novel modules are designed: 1) the spatial and coordinate squeeze-and-excitation fusion attention module that enhances the focus on spatial positions and target channel information; and 2) the efficient multiscale convolutional attention module that can boost the competence to tackle multiscale road information. The validation of the proposed model is conducted using two datasets, CHN6-CUG and DeepGlobe, which are from urban and rural areas, respectively. A comparative analysis with the six commonly used models, including U-Net, PSPNet, DeepLabv3+, HRNet, CoANet, and SegFormer, is conducted. The experimental results reveal that the introduced model achieves mean intersection over union scores of 83.18% and 81.43%, which are averagely improved by 1.93% and 1.02%, respectively, for the two datasets, when compared with the six commonly used models. The outcomes suggest that the introduced model achieves a greater accuracy than the six extensively applied models do.
期刊介绍:
The IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing addresses the growing field of applications in Earth observations and remote sensing, and also provides a venue for the rapidly expanding special issues that are being sponsored by the IEEE Geosciences and Remote Sensing Society. The journal draws upon the experience of the highly successful “IEEE Transactions on Geoscience and Remote Sensing” and provide a complementary medium for the wide range of topics in applied earth observations. The ‘Applications’ areas encompasses the societal benefit areas of the Global Earth Observations Systems of Systems (GEOSS) program. Through deliberations over two years, ministers from 50 countries agreed to identify nine areas where Earth observation could positively impact the quality of life and health of their respective countries. Some of these are areas not traditionally addressed in the IEEE context. These include biodiversity, health and climate. Yet it is the skill sets of IEEE members, in areas such as observations, communications, computers, signal processing, standards and ocean engineering, that form the technical underpinnings of GEOSS. Thus, the Journal attracts a broad range of interests that serves both present members in new ways and expands the IEEE visibility into new areas.