Deep reinforcement learning for irrigation scheduling using high-dimensional sensor feedback

Yuji Saikai, Allan Peake, Karine Chenu
{"title":"Deep reinforcement learning for irrigation scheduling using high-dimensional sensor feedback","authors":"Yuji Saikai, Allan Peake, Karine Chenu","doi":"10.1371/journal.pwat.0000169","DOIUrl":null,"url":null,"abstract":"Deep reinforcement learning has considerable potential to improve irrigation scheduling in many cropping systems by applying adaptive amounts of water based on various measurements over time. The goal is to discover an intelligent decision rule that processes information available to growers and prescribes sensible irrigation amounts for the time steps considered. Due to the technical novelty, however, the research on the technique remains sparse and impractical. To accelerate the progress, the paper proposes a principled framework and actionable procedure that allow researchers to formulate their own optimisation problems and implement solution algorithms based on deep reinforcement learning. The effectiveness of the framework was demonstrated using a case study of irrigated wheat grown in a productive region of Australia where profits were maximised. Specifically, the decision rule takes nine state variable inputs: crop phenological stage, leaf area index, extractable soil water for each of the five top layers, cumulative rainfall and cumulative irrigation. It returns a probabilistic prescription over five candidate irrigation amounts (0, 10, 20, 30 and 40 mm) every day. The production system was simulated at Goondiwindi using the APSIM-Wheat crop model. After training in the learning environment using 1981–2010 weather data, the learned decision rule was tested individually for each year of 2011–2020. The results were compared against the benchmark profits obtained by a conventional rule common in the region. The discovered decision rule prescribed daily irrigation amounts that uniformly improved on the conventional rule for all the testing years, and the largest improvement reached 17% in 2018. The framework is general and applicable to a wide range of cropping systems with realistic optimisation problems.","PeriodicalId":93672,"journal":{"name":"PLOS water","volume":"56 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"PLOS water","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1371/journal.pwat.0000169","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Deep reinforcement learning has considerable potential to improve irrigation scheduling in many cropping systems by applying adaptive amounts of water based on various measurements over time. The goal is to discover an intelligent decision rule that processes information available to growers and prescribes sensible irrigation amounts for the time steps considered. Due to the technical novelty, however, the research on the technique remains sparse and impractical. To accelerate the progress, the paper proposes a principled framework and actionable procedure that allow researchers to formulate their own optimisation problems and implement solution algorithms based on deep reinforcement learning. The effectiveness of the framework was demonstrated using a case study of irrigated wheat grown in a productive region of Australia where profits were maximised. Specifically, the decision rule takes nine state variable inputs: crop phenological stage, leaf area index, extractable soil water for each of the five top layers, cumulative rainfall and cumulative irrigation. It returns a probabilistic prescription over five candidate irrigation amounts (0, 10, 20, 30 and 40 mm) every day. The production system was simulated at Goondiwindi using the APSIM-Wheat crop model. After training in the learning environment using 1981–2010 weather data, the learned decision rule was tested individually for each year of 2011–2020. The results were compared against the benchmark profits obtained by a conventional rule common in the region. The discovered decision rule prescribed daily irrigation amounts that uniformly improved on the conventional rule for all the testing years, and the largest improvement reached 17% in 2018. The framework is general and applicable to a wide range of cropping systems with realistic optimisation problems.
基于高维传感器反馈的灌溉调度深度强化学习
深度强化学习具有相当大的潜力,可以通过根据不同时间的测量值施加自适应水量来改善许多种植系统的灌溉调度。目标是发现一个智能决策规则,该规则可以处理种植者可用的信息,并为所考虑的时间步骤规定合理的灌溉量。然而,由于该技术的新颖性,对该技术的研究仍然很少,而且不切实际。为了加快进展,本文提出了一个原则性框架和可操作的程序,使研究人员能够制定自己的优化问题并实现基于深度强化学习的解决算法。该框架的有效性通过对澳大利亚一个高产地区种植的灌溉小麦的案例研究得到了证明,该地区的利润最大化。具体来说,该决策规则需要9个状态变量输入:作物物候阶段、叶面积指数、5个顶层的可提取土壤水分、累积降雨量和累积灌溉。它每天返回5个候选灌溉量(0、10、20、30和40毫米)的概率处方。利用apsim -小麦作物模型对Goondiwindi的生产系统进行了模拟。在学习环境中使用1981-2010年的天气数据进行训练后,对学习到的决策规则在2011-2020年的每一年进行单独测试。这些结果与该地区常用的常规规则获得的基准利润进行了比较。发现的决策规则规定的日灌溉量在所有测试年份都比常规规则统一提高,2018年最大的提高达到17%。该框架是通用的,适用于广泛的种植系统与现实的优化问题。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信