RAFT: Regularized Adversarial Fine-Tuning to Enhance Deep Reinforcement Learning for Self-Parking

IF 2.2 Q3 ENGINEERING, ELECTRICAL & ELECTRONIC
Alessandro Pighetti;Francesco Bellotti;Riccardo Berta;Andrea Cavallaro;Luca Lazzaroni;Changjae Oh
{"title":"RAFT: Regularized Adversarial Fine-Tuning to Enhance Deep Reinforcement Learning for Self-Parking","authors":"Alessandro Pighetti;Francesco Bellotti;Riccardo Berta;Andrea Cavallaro;Luca Lazzaroni;Changjae Oh","doi":"10.1109/LSENS.2025.3600982","DOIUrl":null,"url":null,"abstract":"Deep reinforcement learning (DRL) is a powerful method for local motion planning in automated driving. However, training of DRL agents is difficult and subject to instability. We propose regularized adversarial fine-tuning (RAFT), an adversarial DRL training framework, and test it in an automated parking (AP) scenario in the car learning to act (CARLA) simulator. Results show that RAFT enhances the performance of a state-of-the-art agent in its original operational design domain (ODD) (static parking, without adversary), by improving its robustness, as evidenced by an increase in all measured metrics. The success rate rises, the mean alignment error shrinks, and the gear reversal rate drops. Notably, we achieved this result not by designing an ad-hoc reward function, but simply by adding a general regularization term to the baseline adversary reward. The results open up new research perspectives for extending the ODD of DRL-based AP to dynamic scenes.","PeriodicalId":13014,"journal":{"name":"IEEE Sensors Letters","volume":"9 9","pages":"1-4"},"PeriodicalIF":2.2000,"publicationDate":"2025-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Sensors Letters","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/11130917/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

Abstract

Deep reinforcement learning (DRL) is a powerful method for local motion planning in automated driving. However, training of DRL agents is difficult and subject to instability. We propose regularized adversarial fine-tuning (RAFT), an adversarial DRL training framework, and test it in an automated parking (AP) scenario in the car learning to act (CARLA) simulator. Results show that RAFT enhances the performance of a state-of-the-art agent in its original operational design domain (ODD) (static parking, without adversary), by improving its robustness, as evidenced by an increase in all measured metrics. The success rate rises, the mean alignment error shrinks, and the gear reversal rate drops. Notably, we achieved this result not by designing an ad-hoc reward function, but simply by adding a general regularization term to the baseline adversary reward. The results open up new research perspectives for extending the ODD of DRL-based AP to dynamic scenes.
RAFT:正则化对抗性微调以增强自动停车的深度强化学习
深度强化学习(DRL)是一种有效的自动驾驶局部运动规划方法。然而,DRL代理的训练是困难的,并且受到不稳定性的影响。我们提出了正则化对抗性微调(RAFT),这是一种对抗性DRL训练框架,并在汽车学习行为(CARLA)模拟器的自动停车(AP)场景中对其进行了测试。结果表明,RAFT通过提高其鲁棒性,增强了最先进智能体在其原始操作设计域(ODD)(静态停车,无对手)中的性能,所有测量指标都有所增加。成功率上升,平均对准误差缩小,齿轮反转率下降。值得注意的是,我们不是通过设计一个特别的奖励函数来实现这个结果的,而是通过简单地向基线对手奖励添加一个通用的正则化项来实现的。研究结果为将基于drl的AP的ODD扩展到动态场景开辟了新的研究视角。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
IEEE Sensors Letters
IEEE Sensors Letters Engineering-Electrical and Electronic Engineering
CiteScore
3.50
自引率
7.10%
发文量
194
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信