Heng Tang , Xiaoping Rui , Jiarui Li , Ninglei Ouyang , Yiheng Xie , Xiaodie Liu , Yiming Bi
{"title":"SLIP-flood: Soft-combination of Swin Transformer and Lightweight Language-Image Pre-training for Flood Images Classification","authors":"Heng Tang , Xiaoping Rui , Jiarui Li , Ninglei Ouyang , Yiheng Xie , Xiaodie Liu , Yiming Bi","doi":"10.1016/j.jag.2025.104543","DOIUrl":null,"url":null,"abstract":"<div><div>Flood monitoring is a complex task involving multimodal data mining and multitask collaboration. In order to leverage the role of multimodal data in flood management, conducting visual-language pretraining (VLP) in the field of flood disaster monitoring and obtaining foundational pretraining models that are suitable for multiple downstream flood-related tasks is an urgent problem that needs to be addressed. This paper introduces SLIP-Flood, an innovative VLP framework supporting flood image classification, image-text retrieval, and auxiliary text classification. To overcome the limitations of existing cross-modal models that rely on small datasets and lack robustness, we have constructed two specialized datasets for the first time: 1) FloodMulS for the Flood Image Classification Model (FICM), and 2) FloodIT for the Flood Text-Image Retrieval Model (FTIRM). Traditional models employ “Hard Categorization Strategy (HC)” for image classification, neglecting the impacts of “Categorization Ambiguity.” To improve performance, we propose a “Soft Categorization Strategy.” Furthermore, traditional models focus on unimodal (image) information, not fully utilizing joint image-text information. We address this with a “Soft Combination” to integrate FICM and FTIRM, termed SCSC. Experimental results show SCSC improves SLIP-Flood’s performance: a 7.62% increase in the F1 score on FICM compared to HC, and a 0.35% increase in FTIRM’s F1 score based on FICM. SLIP-Flood also achieves a maximum recall of 89.24% in image-text retrieval and shows promise in auxiliary flood text classification. Relevant resources are available at https://github.com/muhan-yy/SLIP-Flood.git.</div></div>","PeriodicalId":73423,"journal":{"name":"International journal of applied earth observation and geoinformation : ITC journal","volume":"139 ","pages":"Article 104543"},"PeriodicalIF":7.6000,"publicationDate":"2025-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International journal of applied earth observation and geoinformation : ITC journal","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1569843225001906","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"REMOTE SENSING","Score":null,"Total":0}
引用次数: 0
Abstract
Flood monitoring is a complex task involving multimodal data mining and multitask collaboration. In order to leverage the role of multimodal data in flood management, conducting visual-language pretraining (VLP) in the field of flood disaster monitoring and obtaining foundational pretraining models that are suitable for multiple downstream flood-related tasks is an urgent problem that needs to be addressed. This paper introduces SLIP-Flood, an innovative VLP framework supporting flood image classification, image-text retrieval, and auxiliary text classification. To overcome the limitations of existing cross-modal models that rely on small datasets and lack robustness, we have constructed two specialized datasets for the first time: 1) FloodMulS for the Flood Image Classification Model (FICM), and 2) FloodIT for the Flood Text-Image Retrieval Model (FTIRM). Traditional models employ “Hard Categorization Strategy (HC)” for image classification, neglecting the impacts of “Categorization Ambiguity.” To improve performance, we propose a “Soft Categorization Strategy.” Furthermore, traditional models focus on unimodal (image) information, not fully utilizing joint image-text information. We address this with a “Soft Combination” to integrate FICM and FTIRM, termed SCSC. Experimental results show SCSC improves SLIP-Flood’s performance: a 7.62% increase in the F1 score on FICM compared to HC, and a 0.35% increase in FTIRM’s F1 score based on FICM. SLIP-Flood also achieves a maximum recall of 89.24% in image-text retrieval and shows promise in auxiliary flood text classification. Relevant resources are available at https://github.com/muhan-yy/SLIP-Flood.git.
期刊介绍:
The International Journal of Applied Earth Observation and Geoinformation publishes original papers that utilize earth observation data for natural resource and environmental inventory and management. These data primarily originate from remote sensing platforms, including satellites and aircraft, supplemented by surface and subsurface measurements. Addressing natural resources such as forests, agricultural land, soils, and water, as well as environmental concerns like biodiversity, land degradation, and hazards, the journal explores conceptual and data-driven approaches. It covers geoinformation themes like capturing, databasing, visualization, interpretation, data quality, and spatial uncertainty.