Using Control Theory and Bayesian Reinforcement Learning for Policy Management in Pandemic Situations

Heena Rathore, A. Samant
{"title":"Using Control Theory and Bayesian Reinforcement Learning for Policy Management in Pandemic Situations","authors":"Heena Rathore, A. Samant","doi":"10.1109/ICCWorkshops50388.2021.9473604","DOIUrl":null,"url":null,"abstract":"As engineers and scientists, it is our responsibility to learn lessons from the recent pandemic outbreak and see how public health policies can be effectively managed to reduce the severe loss of lives and minimize the impact on people’s livelihood. Non-pharmaceutical interventions, such as in-place sheltering and social distancing, are typically introduced to slow the spread (flatten the curve) and reverse the growth of the virus. However, such approaches have the unintended consequences of causing economic activities to plummet and bringing local businesses to a standstill, thereby putting millions of jobs at risk. City administrators have generally resorted to an open loop, belief-based decision-making process, thereby struggling to manage (identify and enforce) timely and optimal policies. To overcome this challenge, this position paper explores a systematically designed, feedback-based strategy, to modulate parameters that control suppression and mitigation. Our work leverages advances in Bayesian Reinforcement Learning algorithms and known techniques in control theory, to stabilize and diminish the rate of propagation in pandemic situations. This paper discusses how offline exploitation using pre-trigger data, online exploration using observations from the environment, and a careful orchestration between the two using granular control of multiple on-off control signals can be used to modulate policy enforcement based on established metrics, such as reproduction number.","PeriodicalId":127186,"journal":{"name":"2021 IEEE International Conference on Communications Workshops (ICC Workshops)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2021-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Conference on Communications Workshops (ICC Workshops)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCWorkshops50388.2021.9473604","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

As engineers and scientists, it is our responsibility to learn lessons from the recent pandemic outbreak and see how public health policies can be effectively managed to reduce the severe loss of lives and minimize the impact on people’s livelihood. Non-pharmaceutical interventions, such as in-place sheltering and social distancing, are typically introduced to slow the spread (flatten the curve) and reverse the growth of the virus. However, such approaches have the unintended consequences of causing economic activities to plummet and bringing local businesses to a standstill, thereby putting millions of jobs at risk. City administrators have generally resorted to an open loop, belief-based decision-making process, thereby struggling to manage (identify and enforce) timely and optimal policies. To overcome this challenge, this position paper explores a systematically designed, feedback-based strategy, to modulate parameters that control suppression and mitigation. Our work leverages advances in Bayesian Reinforcement Learning algorithms and known techniques in control theory, to stabilize and diminish the rate of propagation in pandemic situations. This paper discusses how offline exploitation using pre-trigger data, online exploration using observations from the environment, and a careful orchestration between the two using granular control of multiple on-off control signals can be used to modulate policy enforcement based on established metrics, such as reproduction number.
应用控制理论和贝叶斯强化学习在大流行情况下的政策管理
作为工程师和科学家,我们有责任从最近的疫情中吸取教训,研究如何有效管理公共卫生政策,减少严重的生命损失,最大限度地减少对民生的影响。通常采取非药物干预措施,如就地庇护和保持社交距离,以减缓传播(使曲线变平)并扭转病毒的增长。然而,这种做法会产生意想不到的后果,导致经济活动急剧下降,使当地企业陷入停滞,从而使数百万个工作岗位面临风险。城市管理者通常采用开环、基于信念的决策过程,因此难以管理(识别和执行)及时和最佳的政策。为了克服这一挑战,本文探讨了一种系统设计的、基于反馈的策略,以调节控制抑制和缓解的参数。我们的工作利用了贝叶斯强化学习算法的进步和控制理论中的已知技术,以稳定和降低大流行情况下的传播速度。本文讨论了如何使用预触发数据进行离线开发,使用来自环境的观察进行在线探索,以及如何使用多个开关控制信号的粒度控制在两者之间进行精心编排,从而根据已建立的度量(如复制数)来调节策略实施。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信