{"title":"基于强化学习的履带式机器人爬阶鳍运动获取","authors":"Ryosuke Eto, J. Yamakawa","doi":"10.56884/sgwj1011","DOIUrl":null,"url":null,"abstract":"Remotely piloted robots have been expected for disaster operations to prevent secondary disasters to rescuers. The robots are required to have a high performance to overcome obstacles such as debris and bumps. Tracked robots with flipper arms on the front, back, left, and right sides can improve their ability to overcome bumps by changing the flipper angles, and are thus expected to be used as rescue robots. However, the six degrees of freedom of the left and right crawlers and four flipper arms require a high level of skill to maneuver the robot. Therefore, a semi-autonomous control system that automatically controls the flipper arms according to the terrain is expected. In this study, we proposed a method to determine the front and rear flipper angles in step-climbing using reinforcement learning with Double Deep Q Network. The input data were the step height, distance to the step, and current front and rear flipper angles. The outputs were amounts of front and rear flipper angle variations. The behavior of the robot and rewards were calculated using a quasi-static model that considers the slips between the step, floor, and crawler. Positive rewards were given for successful step stepping over steps and negative rewards for unsuccessful steps. Furthermore, the flipper motions at which slippage decreased were obtained by subtracting the sum of the squared values of the slippage rates from the rewards. As the results, it was confirmed that the robot slips less when the body is lifted by the rear flipper than when it runs over along the step shape.","PeriodicalId":447600,"journal":{"name":"Proceedings of the 11th Asia-Pacific Regional Conference of the ISTVS","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Acquisition of Flipper Motion in Step-Climbing of Tracked Robot Using Reinforcement Learning\",\"authors\":\"Ryosuke Eto, J. Yamakawa\",\"doi\":\"10.56884/sgwj1011\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Remotely piloted robots have been expected for disaster operations to prevent secondary disasters to rescuers. The robots are required to have a high performance to overcome obstacles such as debris and bumps. Tracked robots with flipper arms on the front, back, left, and right sides can improve their ability to overcome bumps by changing the flipper angles, and are thus expected to be used as rescue robots. However, the six degrees of freedom of the left and right crawlers and four flipper arms require a high level of skill to maneuver the robot. Therefore, a semi-autonomous control system that automatically controls the flipper arms according to the terrain is expected. In this study, we proposed a method to determine the front and rear flipper angles in step-climbing using reinforcement learning with Double Deep Q Network. The input data were the step height, distance to the step, and current front and rear flipper angles. The outputs were amounts of front and rear flipper angle variations. The behavior of the robot and rewards were calculated using a quasi-static model that considers the slips between the step, floor, and crawler. Positive rewards were given for successful step stepping over steps and negative rewards for unsuccessful steps. Furthermore, the flipper motions at which slippage decreased were obtained by subtracting the sum of the squared values of the slippage rates from the rewards. As the results, it was confirmed that the robot slips less when the body is lifted by the rear flipper than when it runs over along the step shape.\",\"PeriodicalId\":447600,\"journal\":{\"name\":\"Proceedings of the 11th Asia-Pacific Regional Conference of the ISTVS\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-10-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 11th Asia-Pacific Regional Conference of the ISTVS\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.56884/sgwj1011\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 11th Asia-Pacific Regional Conference of the ISTVS","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.56884/sgwj1011","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Acquisition of Flipper Motion in Step-Climbing of Tracked Robot Using Reinforcement Learning
Remotely piloted robots have been expected for disaster operations to prevent secondary disasters to rescuers. The robots are required to have a high performance to overcome obstacles such as debris and bumps. Tracked robots with flipper arms on the front, back, left, and right sides can improve their ability to overcome bumps by changing the flipper angles, and are thus expected to be used as rescue robots. However, the six degrees of freedom of the left and right crawlers and four flipper arms require a high level of skill to maneuver the robot. Therefore, a semi-autonomous control system that automatically controls the flipper arms according to the terrain is expected. In this study, we proposed a method to determine the front and rear flipper angles in step-climbing using reinforcement learning with Double Deep Q Network. The input data were the step height, distance to the step, and current front and rear flipper angles. The outputs were amounts of front and rear flipper angle variations. The behavior of the robot and rewards were calculated using a quasi-static model that considers the slips between the step, floor, and crawler. Positive rewards were given for successful step stepping over steps and negative rewards for unsuccessful steps. Furthermore, the flipper motions at which slippage decreased were obtained by subtracting the sum of the squared values of the slippage rates from the rewards. As the results, it was confirmed that the robot slips less when the body is lifted by the rear flipper than when it runs over along the step shape.