{"title":"Model-Free Deep Reinforcement Learning with Multiple Line-of-Sight Guidance Laws for Autonomous Underwater Vehicles Full-Attitude and Velocity Control","authors":"Chengren Yuan, Changgeng Shuai, Zhanshuo Zhang, Jianguo Ma, Yuan Fang, YuChen Sun","doi":"10.1002/aisy.202400991","DOIUrl":null,"url":null,"abstract":"<p>Autonomous underwater vehicles (AUVs) are increasingly utilized, driving the need for enhanced autonomy. Conventional proportional–integral–derivative (PID) algorithms require frequent control parameter adjustments under varying voyage conditions, which increases operational and experimental costs. To address this issue, a multiple line-of-sight guidance law integrated with a deep reinforcement learning control framework is proposed. This framework enables seamless switching among guidance modes, such as waypoint following, path following, and trajectory tracking, to achieve optimal attitude control. For comprehensive control of roll, pitch, yaw, and longitudinal velocity, an augmented-twin delayed deep deterministic policy gradient (A-TD3) algorithm streamlines the training of the control agent. It enables adaptation to large-range attitude variations using small-scale training data, thereby reducing computational costs for diverse missions. Simulations demonstrate the efficacy of the proposed approach: A-TD3 improves training speed by 30.8% while mitigating issues such as excessive rudder motion, poor operability, and high energy consumption across different missions. The attitude control experiments on the X-AUV prototype validate that A-TD3's control performance with PID method.</p>","PeriodicalId":93858,"journal":{"name":"Advanced intelligent systems (Weinheim an der Bergstrasse, Germany)","volume":"7 8","pages":""},"PeriodicalIF":6.1000,"publicationDate":"2025-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://advanced.onlinelibrary.wiley.com/doi/epdf/10.1002/aisy.202400991","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advanced intelligent systems (Weinheim an der Bergstrasse, Germany)","FirstCategoryId":"1085","ListUrlMain":"https://advanced.onlinelibrary.wiley.com/doi/10.1002/aisy.202400991","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Autonomous underwater vehicles (AUVs) are increasingly utilized, driving the need for enhanced autonomy. Conventional proportional–integral–derivative (PID) algorithms require frequent control parameter adjustments under varying voyage conditions, which increases operational and experimental costs. To address this issue, a multiple line-of-sight guidance law integrated with a deep reinforcement learning control framework is proposed. This framework enables seamless switching among guidance modes, such as waypoint following, path following, and trajectory tracking, to achieve optimal attitude control. For comprehensive control of roll, pitch, yaw, and longitudinal velocity, an augmented-twin delayed deep deterministic policy gradient (A-TD3) algorithm streamlines the training of the control agent. It enables adaptation to large-range attitude variations using small-scale training data, thereby reducing computational costs for diverse missions. Simulations demonstrate the efficacy of the proposed approach: A-TD3 improves training speed by 30.8% while mitigating issues such as excessive rudder motion, poor operability, and high energy consumption across different missions. The attitude control experiments on the X-AUV prototype validate that A-TD3's control performance with PID method.