{"title":"Sampling-based stochastic optimal control with metric interval temporal logic specifications","authors":"Felipe J. Montana, Jun Liu, T. Dodd","doi":"10.1109/CCA.2016.7587911","DOIUrl":null,"url":null,"abstract":"This paper describes a method to find optimal policies for stochastic dynamic systems that maximise the probability of satisfying real-time properties. The method consists of two phases. In the first phase, a coarse abstraction of the original system is created. In each region of the abstraction, a sampling-based algorithm is utilised to compute local policies that allow the system to move between regions. Then, in the second phase, the selection of a policy in each region is obtained by solving a reachability problem on the Cartesian product between the abstraction and a timed automaton representing a real-time specification given as a metric interval temporal logic formula. In contrast to current methods that require a fine abstraction, the proposed method achieves computational tractability by modelling the coarse abstraction of the system as a bounded-parameter Markov decision process (BMDP). Moreover, once the BMDP is created, this can be reused for new specifications assuming the same stochastic system and workspace. The method is demonstrated with an autonomous driving example.","PeriodicalId":301744,"journal":{"name":"2016 IEEE Conference on Control Applications (CCA)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 IEEE Conference on Control Applications (CCA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CCA.2016.7587911","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6
Abstract
This paper describes a method to find optimal policies for stochastic dynamic systems that maximise the probability of satisfying real-time properties. The method consists of two phases. In the first phase, a coarse abstraction of the original system is created. In each region of the abstraction, a sampling-based algorithm is utilised to compute local policies that allow the system to move between regions. Then, in the second phase, the selection of a policy in each region is obtained by solving a reachability problem on the Cartesian product between the abstraction and a timed automaton representing a real-time specification given as a metric interval temporal logic formula. In contrast to current methods that require a fine abstraction, the proposed method achieves computational tractability by modelling the coarse abstraction of the system as a bounded-parameter Markov decision process (BMDP). Moreover, once the BMDP is created, this can be reused for new specifications assuming the same stochastic system and workspace. The method is demonstrated with an autonomous driving example.