{"title":"MMF-RNN: A Multimodal Fusion Model for Precipitation Nowcasting Using Radar and Ground Station Data","authors":"Qian Liu;Yongxiang Xiao;Yaocheng Gui;Guilan Dai;Haoran Li;Xu Zhou;Aiai Ren;Guoqiang Zhou;Jun Shen","doi":"10.1109/TGRS.2025.3528423","DOIUrl":null,"url":null,"abstract":"Precipitation nowcasting is crucial for economic development and social life. Numerous deep learning models have recently been developed and have achieved better results than traditional extrapolation models. However, they mainly focus on improving model architectures, ignoring the impact of error accumulation and data inconsistency. This article proposes a multimodal fusion model named multimodal fusion recurrent neural network (MMF-RNN) for precipitation prediction. Specifically, we use a dual-branch encoder to extract features from radar and ground station data and then fuse them effectively through attention mechanisms and multimodal loss (ML). To address the error accumulation problem, we propose a block-based dynamic weighted loss (BDWLoss) that enables the model to focus more on hard-to-predict areas during training to reduce error accumulation. Based on BDWLoss, we propose an ML that encourages the model to maintain consistency between single-modal and fused multimodal features. In addition, MMF-RNN is compatible with various RNN models such as ConvLSTM, PredRNN, PredRNN++, and MIM. The experimental results on the RAIN-F dataset demonstrate that MMF-RNN outperforms both the single-modal model MS-RNN and the multimodal model MM-RNN. In particular, MMF-RNN achieves significant improvement in predicting heavy precipitation. Compared to MM-PredRNN++, MMF-PredRNN++ shows marked improvements across various performance metrics, with critical success index (CSI) (<inline-formula> <tex-math>$R\\geq 5 $ </tex-math></inline-formula>) and Heidke skill score (HSS) (<inline-formula> <tex-math>$R\\geq 5$ </tex-math></inline-formula>) increasing by 58.08% and 48.55%, respectively, and CSI (<inline-formula> <tex-math>$R\\geq 10$ </tex-math></inline-formula>) and HSS (<inline-formula> <tex-math>$R\\geq 10$ </tex-math></inline-formula>) showing more pronounced gains. These advancements are facilitated not only by the proposed architectural innovations but also by sample weighting, which collectively contribute to superior performance on imbalanced precipitation datasets.","PeriodicalId":13213,"journal":{"name":"IEEE Transactions on Geoscience and Remote Sensing","volume":"63 ","pages":"1-16"},"PeriodicalIF":8.6000,"publicationDate":"2025-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Geoscience and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10845122/","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Precipitation nowcasting is crucial for economic development and social life. Numerous deep learning models have recently been developed and have achieved better results than traditional extrapolation models. However, they mainly focus on improving model architectures, ignoring the impact of error accumulation and data inconsistency. This article proposes a multimodal fusion model named multimodal fusion recurrent neural network (MMF-RNN) for precipitation prediction. Specifically, we use a dual-branch encoder to extract features from radar and ground station data and then fuse them effectively through attention mechanisms and multimodal loss (ML). To address the error accumulation problem, we propose a block-based dynamic weighted loss (BDWLoss) that enables the model to focus more on hard-to-predict areas during training to reduce error accumulation. Based on BDWLoss, we propose an ML that encourages the model to maintain consistency between single-modal and fused multimodal features. In addition, MMF-RNN is compatible with various RNN models such as ConvLSTM, PredRNN, PredRNN++, and MIM. The experimental results on the RAIN-F dataset demonstrate that MMF-RNN outperforms both the single-modal model MS-RNN and the multimodal model MM-RNN. In particular, MMF-RNN achieves significant improvement in predicting heavy precipitation. Compared to MM-PredRNN++, MMF-PredRNN++ shows marked improvements across various performance metrics, with critical success index (CSI) ($R\geq 5 $ ) and Heidke skill score (HSS) ($R\geq 5$ ) increasing by 58.08% and 48.55%, respectively, and CSI ($R\geq 10$ ) and HSS ($R\geq 10$ ) showing more pronounced gains. These advancements are facilitated not only by the proposed architectural innovations but also by sample weighting, which collectively contribute to superior performance on imbalanced precipitation datasets.
降水临近预报对经济发展和社会生活至关重要。最近开发了许多深度学习模型,并取得了比传统外推模型更好的结果。然而,他们主要关注于改进模型架构,忽略了错误积累和数据不一致的影响。提出了一种用于降水预报的多模态融合递归神经网络(MMF-RNN)模型。具体来说,我们使用双分支编码器从雷达和地面站数据中提取特征,然后通过注意机制和多模态损失(ML)有效地融合它们。为了解决误差积累问题,我们提出了一种基于块的动态加权损失(BDWLoss),使模型在训练过程中更多地关注难以预测的区域,以减少误差积累。基于BDWLoss,我们提出了一种机器学习,鼓励模型保持单模态和融合的多模态特征之间的一致性。此外,MMF-RNN还兼容ConvLSTM、PredRNN、PredRNN++、MIM等多种RNN模型。在RAIN-F数据集上的实验结果表明,MMF-RNN优于单模态模型MS-RNN和多模态模型MM-RNN。特别是MMF-RNN在预测强降水方面取得了显著的进步。与MM-PredRNN++相比,MMF-PredRNN++在各种性能指标上都有显著改善,关键成功指数(CSI) ($R\geq 5 $)和Heidke技能分数(HSS) ($R\geq 5$)提高了58.08% and 48.55%, respectively, and CSI ( $R\geq 10$ ) and HSS ( $R\geq 10$ ) showing more pronounced gains. These advancements are facilitated not only by the proposed architectural innovations but also by sample weighting, which collectively contribute to superior performance on imbalanced precipitation datasets.
期刊介绍:
IEEE Transactions on Geoscience and Remote Sensing (TGRS) is a monthly publication that focuses on the theory, concepts, and techniques of science and engineering as applied to sensing the land, oceans, atmosphere, and space; and the processing, interpretation, and dissemination of this information.