Hozefa Jesawada;Antonio Acernese;Davide Del Vecchio;Giovanni Russo;Carmen Del Vecchio
{"title":"DR-PETS: Learning-Based Control With Planning in Adversarial Environments","authors":"Hozefa Jesawada;Antonio Acernese;Davide Del Vecchio;Giovanni Russo;Carmen Del Vecchio","doi":"10.1109/LCSYS.2025.3587053","DOIUrl":null,"url":null,"abstract":"The probabilistic ensembles with trajectory sampling (PETS) algorithm is a recognized baseline among model-based reinforcement learning (MBRL) methods. PETS incorporates planning and handles uncertainty using ensemble-based probabilistic models. However, no formal robustness guarantees against epistemic uncertainty exist for PETS. Providing such guarantees is a key enabler for reliable real-world deployment. To address this gap, we propose a distributionally robust extension of PETS, called DR-PETS. We formalize model uncertainty using a distributional ambiguity set and optimize the worst-case expected return. We derive a tractable convex approximation of the resulting min-max planning problem, which integrates seamlessly into PETS’s planning loop as a regularized objective. Experiments on pendulum and cart-pole environments show that DR-PETS certifies robustness against adversarial parameter perturbations, achieving consistent performance in worst-case scenarios where PETS deteriorates.","PeriodicalId":37235,"journal":{"name":"IEEE Control Systems Letters","volume":"9 ","pages":"2006-2011"},"PeriodicalIF":2.0000,"publicationDate":"2025-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Control Systems Letters","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/11075840/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
The probabilistic ensembles with trajectory sampling (PETS) algorithm is a recognized baseline among model-based reinforcement learning (MBRL) methods. PETS incorporates planning and handles uncertainty using ensemble-based probabilistic models. However, no formal robustness guarantees against epistemic uncertainty exist for PETS. Providing such guarantees is a key enabler for reliable real-world deployment. To address this gap, we propose a distributionally robust extension of PETS, called DR-PETS. We formalize model uncertainty using a distributional ambiguity set and optimize the worst-case expected return. We derive a tractable convex approximation of the resulting min-max planning problem, which integrates seamlessly into PETS’s planning loop as a regularized objective. Experiments on pendulum and cart-pole environments show that DR-PETS certifies robustness against adversarial parameter perturbations, achieving consistent performance in worst-case scenarios where PETS deteriorates.