Ruiyang Sun , Xin Su , Qiangqiang Yuan , Hongzan Jiao , Jiang He , Li Zheng
{"title":"Urban region function classification via fusing optical imagery and social media data: A spatio-temporal Transformer interaction approach","authors":"Ruiyang Sun , Xin Su , Qiangqiang Yuan , Hongzan Jiao , Jiang He , Li Zheng","doi":"10.1016/j.inffus.2025.103140","DOIUrl":null,"url":null,"abstract":"<div><div>Urban management and planning can benefit from the classification of urban functional areas. Existing researches have demonstrated that remote sensing data can provide essential urban surface spatial information for the identification of urban functional areas, and human activities may characterize the dynamic aspects connected to social and economic temporal information in various urban functional regions. However, current methods lack explicit consideration of the mutual properties of spatial and temporal modality, resulting in suboptimal interaction performance. To address the issue, we propose a novel fusion method of spatio-temporal Transformer of remote sensing and social media data for urban region function classification. We design the Multi-scale Vision Transformer (MultiViT) to extract the multi-scale features of optical image data and Convolutional Transformer (ConvTransformer) to obtain the multi-scale temporal scale features of social media time series data. For multi-modal fusion, we create a crucial spatio-temporal fusion path based on self-attention, using the different modalities semantic information regarded as useful priori information. By the supervised and distilled loss function, the merging of the two sub-networks and the main network is taken into account during training. Extensive experiments on public datasets have demonstrated the favorable performance of our spatio-temporal Transformer interaction approach in merging remote sensing and social media data for urban region function classification. The code will be available at <span><span>https://github.com/Ruiyang-Sun/Spatio-temporal-Transformer-for-urfc</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"121 ","pages":"Article 103140"},"PeriodicalIF":14.7000,"publicationDate":"2025-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525002131","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Urban management and planning can benefit from the classification of urban functional areas. Existing researches have demonstrated that remote sensing data can provide essential urban surface spatial information for the identification of urban functional areas, and human activities may characterize the dynamic aspects connected to social and economic temporal information in various urban functional regions. However, current methods lack explicit consideration of the mutual properties of spatial and temporal modality, resulting in suboptimal interaction performance. To address the issue, we propose a novel fusion method of spatio-temporal Transformer of remote sensing and social media data for urban region function classification. We design the Multi-scale Vision Transformer (MultiViT) to extract the multi-scale features of optical image data and Convolutional Transformer (ConvTransformer) to obtain the multi-scale temporal scale features of social media time series data. For multi-modal fusion, we create a crucial spatio-temporal fusion path based on self-attention, using the different modalities semantic information regarded as useful priori information. By the supervised and distilled loss function, the merging of the two sub-networks and the main network is taken into account during training. Extensive experiments on public datasets have demonstrated the favorable performance of our spatio-temporal Transformer interaction approach in merging remote sensing and social media data for urban region function classification. The code will be available at https://github.com/Ruiyang-Sun/Spatio-temporal-Transformer-for-urfc.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.