{"title":"Toward Improving the Distributional Robustness of Risk-Aware Controllers in Learning-Enabled Environments","authors":"A. Hakobyan, Insoon Yang","doi":"10.1109/CDC45484.2021.9682981","DOIUrl":null,"url":null,"abstract":"This paper is concerned with designing a risk-aware controller in an unknown and dynamic environment. In our method, the evolution of the environment state is learned using observational data via Gaussian process regression (GPR). Unfortunately, these learning results provide imperfect distribution information about the environment. To address such distribution errors, we propose a risk-constrained model predictive control (MPC) method that exploits techniques from modern distributionally robust optimization (DRO). To resolve the infinite dimensionality issue inherent in DRO, we derive a tractable semidefinite programming (SDP) problem that upper-bounds the original MPC problem. Furthermore, the SDP problem is reduced to a quadratic program when the constraint function has a decomposable form. The performance and the utility of our method are demonstrated through an autonomous driving problem, and the results show that our controller preserves safety despite errors in learning the behaviors of surrounding vehicles.","PeriodicalId":229089,"journal":{"name":"2021 60th IEEE Conference on Decision and Control (CDC)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 60th IEEE Conference on Decision and Control (CDC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CDC45484.2021.9682981","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
This paper is concerned with designing a risk-aware controller in an unknown and dynamic environment. In our method, the evolution of the environment state is learned using observational data via Gaussian process regression (GPR). Unfortunately, these learning results provide imperfect distribution information about the environment. To address such distribution errors, we propose a risk-constrained model predictive control (MPC) method that exploits techniques from modern distributionally robust optimization (DRO). To resolve the infinite dimensionality issue inherent in DRO, we derive a tractable semidefinite programming (SDP) problem that upper-bounds the original MPC problem. Furthermore, the SDP problem is reduced to a quadratic program when the constraint function has a decomposable form. The performance and the utility of our method are demonstrated through an autonomous driving problem, and the results show that our controller preserves safety despite errors in learning the behaviors of surrounding vehicles.