Adaptive fusion of multi-modal remote sensing data for optimal sub-field crop yield prediction

IF 11.1 1区 地球科学 Q1 ENVIRONMENTAL SCIENCES
Francisco Mena , Deepak Pathak , Hiba Najjar , Cristhian Sanchez , Patrick Helber , Benjamin Bischke , Peter Habelitz , Miro Miranda , Jayanth Siddamsetty , Marlon Nuske , Marcela Charfuelan , Diego Arenas , Michaela Vollmer , Andreas Dengel
{"title":"Adaptive fusion of multi-modal remote sensing data for optimal sub-field crop yield prediction","authors":"Francisco Mena ,&nbsp;Deepak Pathak ,&nbsp;Hiba Najjar ,&nbsp;Cristhian Sanchez ,&nbsp;Patrick Helber ,&nbsp;Benjamin Bischke ,&nbsp;Peter Habelitz ,&nbsp;Miro Miranda ,&nbsp;Jayanth Siddamsetty ,&nbsp;Marlon Nuske ,&nbsp;Marcela Charfuelan ,&nbsp;Diego Arenas ,&nbsp;Michaela Vollmer ,&nbsp;Andreas Dengel","doi":"10.1016/j.rse.2024.114547","DOIUrl":null,"url":null,"abstract":"<div><div>Accurate crop yield prediction is of utmost importance for informed decision-making in agriculture, aiding farmers, industry stakeholders, and policymakers in optimizing agricultural practices. However, this task is complex and depends on multiple factors, such as environmental conditions, soil properties, and management practices. Leveraging Remote Sensing (RS) technologies, multi-modal data from diverse global data sources can be collected to enhance predictive model accuracy. However, combining heterogeneous RS data poses a fusion challenge, like identifying the specific contribution of each modality in the predictive task. In this paper, we present a novel multi-modal learning approach to predict crop yield for different crops (soybean, wheat, rapeseed) and regions (Argentina, Uruguay, and Germany). Our multi-modal input data includes multi-spectral optical images from Sentinel-2 satellites and weather data as dynamic features during the crop growing season, complemented by static features like soil properties and topographic information. To effectively fuse the multi-modal data, we introduce a Multi-modal Gated Fusion (MMGF) model, comprising dedicated modality-encoders and a Gated Unit (GU) module. The modality-encoders handle the heterogeneity of data sources with varying temporal resolutions by learning a modality-specific representation. These representations are adaptively fused via a weighted sum. The <em>fusion</em> weights are computed for each sample by the GU using a concatenation of the multi-modal representations. The MMGF model is trained at sub-field level with 10 m resolution pixels. Our evaluations show that the MMGF outperforms conventional models on the same task, achieving the best results by incorporating all the data sources, unlike the usual fusion results in the literature. For Argentina, the MMGF model achieves an <span><math><msup><mrow><mi>R</mi></mrow><mrow><mn>2</mn></mrow></msup></math></span> value of 0.68 at sub-field yield prediction, while at the field level evaluation (comparing field averages), it reaches around 0.80 across different countries. The GU module learned different weights based on the country and crop-type, aligning with the variable significance of each data source to the prediction task. This novel method has proven its effectiveness in enhancing the accuracy of the challenging sub-field crop yield prediction. Our investigation indicates that the gated fusion approach promises a significant advancement in the field of agriculture and precision farming.</div></div>","PeriodicalId":417,"journal":{"name":"Remote Sensing of Environment","volume":"318 ","pages":"Article 114547"},"PeriodicalIF":11.1000,"publicationDate":"2024-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Remote Sensing of Environment","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S003442572400573X","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENVIRONMENTAL SCIENCES","Score":null,"Total":0}
引用次数: 0

Abstract

Accurate crop yield prediction is of utmost importance for informed decision-making in agriculture, aiding farmers, industry stakeholders, and policymakers in optimizing agricultural practices. However, this task is complex and depends on multiple factors, such as environmental conditions, soil properties, and management practices. Leveraging Remote Sensing (RS) technologies, multi-modal data from diverse global data sources can be collected to enhance predictive model accuracy. However, combining heterogeneous RS data poses a fusion challenge, like identifying the specific contribution of each modality in the predictive task. In this paper, we present a novel multi-modal learning approach to predict crop yield for different crops (soybean, wheat, rapeseed) and regions (Argentina, Uruguay, and Germany). Our multi-modal input data includes multi-spectral optical images from Sentinel-2 satellites and weather data as dynamic features during the crop growing season, complemented by static features like soil properties and topographic information. To effectively fuse the multi-modal data, we introduce a Multi-modal Gated Fusion (MMGF) model, comprising dedicated modality-encoders and a Gated Unit (GU) module. The modality-encoders handle the heterogeneity of data sources with varying temporal resolutions by learning a modality-specific representation. These representations are adaptively fused via a weighted sum. The fusion weights are computed for each sample by the GU using a concatenation of the multi-modal representations. The MMGF model is trained at sub-field level with 10 m resolution pixels. Our evaluations show that the MMGF outperforms conventional models on the same task, achieving the best results by incorporating all the data sources, unlike the usual fusion results in the literature. For Argentina, the MMGF model achieves an R2 value of 0.68 at sub-field yield prediction, while at the field level evaluation (comparing field averages), it reaches around 0.80 across different countries. The GU module learned different weights based on the country and crop-type, aligning with the variable significance of each data source to the prediction task. This novel method has proven its effectiveness in enhancing the accuracy of the challenging sub-field crop yield prediction. Our investigation indicates that the gated fusion approach promises a significant advancement in the field of agriculture and precision farming.
多模态遥感数据自适应融合优化子田作物产量预测
准确的作物产量预测对农业决策至关重要,可帮助农民、行业利益相关者和政策制定者优化农业实践。然而,这项任务十分复杂,取决于环境条件、土壤特性和管理方法等多种因素。利用遥感(RS)技术,可以从不同的全球数据源收集多模式数据,从而提高预测模型的准确性。然而,结合异构 RS 数据会带来融合方面的挑战,如确定每种模式在预测任务中的具体贡献。在本文中,我们提出了一种新颖的多模态学习方法,用于预测不同作物(大豆、小麦、油菜籽)和地区(阿根廷、乌拉圭和德国)的作物产量。我们的多模态输入数据包括来自 Sentinel-2 卫星的多光谱光学图像和作物生长季节的气象数据作为动态特征,并辅以土壤特性和地形信息等静态特征。为有效融合多模态数据,我们引入了多模态门控融合(MMGF)模型,该模型由专用模态编码器和门控单元(GU)模块组成。模态编码器通过学习特定模态的表示来处理具有不同时间分辨率的数据源的异质性。这些表征通过加权和进行自适应融合。融合权重由 GU 使用多模态表征的连接为每个样本计算。MMGF 模型是在 10 米分辨率像素的子场级别进行训练的。我们的评估结果表明,在同一任务中,MMGF 的表现优于传统模型,它通过整合所有数据源取得了最佳结果,这与文献中通常的融合结果不同。在阿根廷,MMGF 模型在分田产量预测方面的 R2R2 值达到 0.68,而在田间水平评估(比较田间平均值)方面,不同国家的 R2R2 值达到 0.80 左右。基于国家和作物类型,GU 模块学习了不同的权重,与每个数据源对预测任务的变量重要性相一致。事实证明,这种新方法能有效提高具有挑战性的分田作物产量预测的准确性。我们的调查表明,门控融合方法有望在农业和精准农业领域取得重大进展。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Remote Sensing of Environment
Remote Sensing of Environment 环境科学-成像科学与照相技术
CiteScore
25.10
自引率
8.90%
发文量
455
审稿时长
53 days
期刊介绍: Remote Sensing of Environment (RSE) serves the Earth observation community by disseminating results on the theory, science, applications, and technology that contribute to advancing the field of remote sensing. With a thoroughly interdisciplinary approach, RSE encompasses terrestrial, oceanic, and atmospheric sensing. The journal emphasizes biophysical and quantitative approaches to remote sensing at local to global scales, covering a diverse range of applications and techniques. RSE serves as a vital platform for the exchange of knowledge and advancements in the dynamic field of remote sensing.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信