Finite horizon partially observable semi-Markov decision processes under risk probability criteria

IF 0.8 4区 管理学 Q4 OPERATIONS RESEARCH & MANAGEMENT SCIENCE
Xin Wen , Xianping Guo , Li Xia
{"title":"Finite horizon partially observable semi-Markov decision processes under risk probability criteria","authors":"Xin Wen ,&nbsp;Xianping Guo ,&nbsp;Li Xia","doi":"10.1016/j.orl.2024.107187","DOIUrl":null,"url":null,"abstract":"<div><div>This paper deals with a risk probability minimization problem for finite horizon partially observable semi-Markov decision processes, which are the fairly most general models for stochastic dynamic systems. In contrast to the expected discounted and average criteria, the optimality investigated in this paper is to minimize the probability that the accumulated rewards do not reach a prescribed profit level at the finite terminal stage. First, the state space is augmented as the joint conditional distribution of the current unobserved state and the remaining profit goal. We introduce a class of policies depending on observable histories and a class of Markov policies including observable process with the joint conditional distribution. Then under mild assumptions, we prove that the value function is the unique solution to the optimality equation for the probability criterion by using iteration techniques. The existence of (<em>ϵ</em>-)optimal Markov policy for this problem is established. Finally, we use a bandit problem with the probability criterion to demonstrate our main results in which an effective algorithm and the corresponding numerical calculation are given for the semi-Markov model. Moreover, for the case of reduction to the discrete-time Markov model, we derive a concise solution.</div></div>","PeriodicalId":54682,"journal":{"name":"Operations Research Letters","volume":"57 ","pages":"Article 107187"},"PeriodicalIF":0.8000,"publicationDate":"2024-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Operations Research Letters","FirstCategoryId":"91","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167637724001238","RegionNum":4,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"OPERATIONS RESEARCH & MANAGEMENT SCIENCE","Score":null,"Total":0}
引用次数: 0

Abstract

This paper deals with a risk probability minimization problem for finite horizon partially observable semi-Markov decision processes, which are the fairly most general models for stochastic dynamic systems. In contrast to the expected discounted and average criteria, the optimality investigated in this paper is to minimize the probability that the accumulated rewards do not reach a prescribed profit level at the finite terminal stage. First, the state space is augmented as the joint conditional distribution of the current unobserved state and the remaining profit goal. We introduce a class of policies depending on observable histories and a class of Markov policies including observable process with the joint conditional distribution. Then under mild assumptions, we prove that the value function is the unique solution to the optimality equation for the probability criterion by using iteration techniques. The existence of (ϵ-)optimal Markov policy for this problem is established. Finally, we use a bandit problem with the probability criterion to demonstrate our main results in which an effective algorithm and the corresponding numerical calculation are given for the semi-Markov model. Moreover, for the case of reduction to the discrete-time Markov model, we derive a concise solution.
风险概率标准下的有限视界部分可观测半马尔可夫决策过程
本文研究的是有限视距部分可观测半马尔可夫决策过程的风险概率最小化问题,半马尔可夫决策过程是随机动态系统的最一般模型。与预期贴现和平均准则不同,本文研究的最优性是最小化累积奖励在有限终结阶段达不到规定利润水平的概率。首先,状态空间被扩展为当前未观测状态和剩余利润目标的联合条件分布。我们引入了一类取决于可观测历史的策略,以及一类包含联合条件分布的可观测过程的马尔可夫策略。然后,在温和的假设条件下,我们利用迭代技术证明价值函数是概率准则最优方程的唯一解。针对这个问题,我们确定了 (ϵ-)optimal Markov policy 的存在。最后,我们用一个带有概率准则的强盗问题来证明我们的主要成果,其中给出了半马尔可夫模型的有效算法和相应的数值计算。此外,对于还原到离散时间马尔可夫模型的情况,我们还得出了一个简明的解决方案。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Operations Research Letters
Operations Research Letters 管理科学-运筹学与管理科学
CiteScore
2.10
自引率
9.10%
发文量
111
审稿时长
83 days
期刊介绍: Operations Research Letters is committed to the rapid review and fast publication of short articles on all aspects of operations research and analytics. Apart from a limitation to eight journal pages, quality, originality, relevance and clarity are the only criteria for selecting the papers to be published. ORL covers the broad field of optimization, stochastic models and game theory. Specific areas of interest include networks, routing, location, queueing, scheduling, inventory, reliability, and financial engineering. We wish to explore interfaces with other fields such as life sciences and health care, artificial intelligence and machine learning, energy distribution, and computational social sciences and humanities. Our traditional strength is in methodology, including theory, modelling, algorithms and computational studies. We also welcome novel applications and concise literature reviews.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信