{"title":"RAFT:正则化对抗性微调以增强自动停车的深度强化学习","authors":"Alessandro Pighetti;Francesco Bellotti;Riccardo Berta;Andrea Cavallaro;Luca Lazzaroni;Changjae Oh","doi":"10.1109/LSENS.2025.3600982","DOIUrl":null,"url":null,"abstract":"Deep reinforcement learning (DRL) is a powerful method for local motion planning in automated driving. However, training of DRL agents is difficult and subject to instability. We propose regularized adversarial fine-tuning (RAFT), an adversarial DRL training framework, and test it in an automated parking (AP) scenario in the car learning to act (CARLA) simulator. Results show that RAFT enhances the performance of a state-of-the-art agent in its original operational design domain (ODD) (static parking, without adversary), by improving its robustness, as evidenced by an increase in all measured metrics. The success rate rises, the mean alignment error shrinks, and the gear reversal rate drops. Notably, we achieved this result not by designing an ad-hoc reward function, but simply by adding a general regularization term to the baseline adversary reward. The results open up new research perspectives for extending the ODD of DRL-based AP to dynamic scenes.","PeriodicalId":13014,"journal":{"name":"IEEE Sensors Letters","volume":"9 9","pages":"1-4"},"PeriodicalIF":2.2000,"publicationDate":"2025-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"RAFT: Regularized Adversarial Fine-Tuning to Enhance Deep Reinforcement Learning for Self-Parking\",\"authors\":\"Alessandro Pighetti;Francesco Bellotti;Riccardo Berta;Andrea Cavallaro;Luca Lazzaroni;Changjae Oh\",\"doi\":\"10.1109/LSENS.2025.3600982\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Deep reinforcement learning (DRL) is a powerful method for local motion planning in automated driving. However, training of DRL agents is difficult and subject to instability. We propose regularized adversarial fine-tuning (RAFT), an adversarial DRL training framework, and test it in an automated parking (AP) scenario in the car learning to act (CARLA) simulator. Results show that RAFT enhances the performance of a state-of-the-art agent in its original operational design domain (ODD) (static parking, without adversary), by improving its robustness, as evidenced by an increase in all measured metrics. The success rate rises, the mean alignment error shrinks, and the gear reversal rate drops. Notably, we achieved this result not by designing an ad-hoc reward function, but simply by adding a general regularization term to the baseline adversary reward. The results open up new research perspectives for extending the ODD of DRL-based AP to dynamic scenes.\",\"PeriodicalId\":13014,\"journal\":{\"name\":\"IEEE Sensors Letters\",\"volume\":\"9 9\",\"pages\":\"1-4\"},\"PeriodicalIF\":2.2000,\"publicationDate\":\"2025-08-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Sensors Letters\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/11130917/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Sensors Letters","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/11130917/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
RAFT: Regularized Adversarial Fine-Tuning to Enhance Deep Reinforcement Learning for Self-Parking
Deep reinforcement learning (DRL) is a powerful method for local motion planning in automated driving. However, training of DRL agents is difficult and subject to instability. We propose regularized adversarial fine-tuning (RAFT), an adversarial DRL training framework, and test it in an automated parking (AP) scenario in the car learning to act (CARLA) simulator. Results show that RAFT enhances the performance of a state-of-the-art agent in its original operational design domain (ODD) (static parking, without adversary), by improving its robustness, as evidenced by an increase in all measured metrics. The success rate rises, the mean alignment error shrinks, and the gear reversal rate drops. Notably, we achieved this result not by designing an ad-hoc reward function, but simply by adding a general regularization term to the baseline adversary reward. The results open up new research perspectives for extending the ODD of DRL-based AP to dynamic scenes.