{"title":"Sarsa视觉伺服增益调谐:在机械臂上的应用","authors":"Jie Liu, Yang Zhou, Jian Gao, Weisheng Yan","doi":"10.1145/3598151.3598169","DOIUrl":null,"url":null,"abstract":"This paper investigates a Sarsa-based visual servoing control gain tuning method and the application on a manipulator. For a typical visual servo controller, fixed control gains will not provide the best performance. Therefore, state action reward state action (SARSA) algorithm, one of learning-based methods from reinforcement learning (RL), is introduced to select control gains in every control step. The norm of the visual error is used to define the state space. The positive gain of the controller is discretized as the actions. A reward function is defined to evaluate the performance of every action. Both a numerical test and a robot experiment are carried out to validate the presented algorithm.","PeriodicalId":398644,"journal":{"name":"Proceedings of the 2023 3rd International Conference on Robotics and Control Engineering","volume":"240 1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Visual Servoing Gain Tuning by Sarsa: an Application with a Manipulator\",\"authors\":\"Jie Liu, Yang Zhou, Jian Gao, Weisheng Yan\",\"doi\":\"10.1145/3598151.3598169\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper investigates a Sarsa-based visual servoing control gain tuning method and the application on a manipulator. For a typical visual servo controller, fixed control gains will not provide the best performance. Therefore, state action reward state action (SARSA) algorithm, one of learning-based methods from reinforcement learning (RL), is introduced to select control gains in every control step. The norm of the visual error is used to define the state space. The positive gain of the controller is discretized as the actions. A reward function is defined to evaluate the performance of every action. Both a numerical test and a robot experiment are carried out to validate the presented algorithm.\",\"PeriodicalId\":398644,\"journal\":{\"name\":\"Proceedings of the 2023 3rd International Conference on Robotics and Control Engineering\",\"volume\":\"240 1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-05-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2023 3rd International Conference on Robotics and Control Engineering\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3598151.3598169\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2023 3rd International Conference on Robotics and Control Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3598151.3598169","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Visual Servoing Gain Tuning by Sarsa: an Application with a Manipulator
This paper investigates a Sarsa-based visual servoing control gain tuning method and the application on a manipulator. For a typical visual servo controller, fixed control gains will not provide the best performance. Therefore, state action reward state action (SARSA) algorithm, one of learning-based methods from reinforcement learning (RL), is introduced to select control gains in every control step. The norm of the visual error is used to define the state space. The positive gain of the controller is discretized as the actions. A reward function is defined to evaluate the performance of every action. Both a numerical test and a robot experiment are carried out to validate the presented algorithm.