Sampling-based stochastic optimal control with metric interval temporal logic specifications

Felipe J. Montana, Jun Liu, T. Dodd
{"title":"Sampling-based stochastic optimal control with metric interval temporal logic specifications","authors":"Felipe J. Montana, Jun Liu, T. Dodd","doi":"10.1109/CCA.2016.7587911","DOIUrl":null,"url":null,"abstract":"This paper describes a method to find optimal policies for stochastic dynamic systems that maximise the probability of satisfying real-time properties. The method consists of two phases. In the first phase, a coarse abstraction of the original system is created. In each region of the abstraction, a sampling-based algorithm is utilised to compute local policies that allow the system to move between regions. Then, in the second phase, the selection of a policy in each region is obtained by solving a reachability problem on the Cartesian product between the abstraction and a timed automaton representing a real-time specification given as a metric interval temporal logic formula. In contrast to current methods that require a fine abstraction, the proposed method achieves computational tractability by modelling the coarse abstraction of the system as a bounded-parameter Markov decision process (BMDP). Moreover, once the BMDP is created, this can be reused for new specifications assuming the same stochastic system and workspace. The method is demonstrated with an autonomous driving example.","PeriodicalId":301744,"journal":{"name":"2016 IEEE Conference on Control Applications (CCA)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 IEEE Conference on Control Applications (CCA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CCA.2016.7587911","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6

Abstract

This paper describes a method to find optimal policies for stochastic dynamic systems that maximise the probability of satisfying real-time properties. The method consists of two phases. In the first phase, a coarse abstraction of the original system is created. In each region of the abstraction, a sampling-based algorithm is utilised to compute local policies that allow the system to move between regions. Then, in the second phase, the selection of a policy in each region is obtained by solving a reachability problem on the Cartesian product between the abstraction and a timed automaton representing a real-time specification given as a metric interval temporal logic formula. In contrast to current methods that require a fine abstraction, the proposed method achieves computational tractability by modelling the coarse abstraction of the system as a bounded-parameter Markov decision process (BMDP). Moreover, once the BMDP is created, this can be reused for new specifications assuming the same stochastic system and workspace. The method is demonstrated with an autonomous driving example.
基于抽样的度量区间时间逻辑规范随机最优控制
本文描述了一种寻找随机动态系统的最优策略的方法,该策略使满足实时特性的概率最大化。该方法包括两个阶段。在第一阶段,创建原始系统的粗略抽象。在抽象的每个区域中,使用基于采样的算法来计算允许系统在区域之间移动的本地策略。然后,在第二阶段,通过求解抽象和时间自动机之间笛卡尔积的可达性问题,得到每个区域的策略选择,时间自动机表示以度量区间时间逻辑公式给出的实时规范。与当前需要精细抽象的方法相比,该方法通过将系统的粗抽象建模为有界参数马尔可夫决策过程(BMDP)来实现计算可追溯性。此外,一旦创建了BMDP,就可以在假设相同的随机系统和工作空间的新规范中重用它。通过一个自动驾驶实例对该方法进行了验证。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信