{"title":"基于单目摄像机的关节机器人鲁棒少镜头姿态估计及深度学习关键点检测","authors":"Jens Lambrecht","doi":"10.1109/RITAPP.2019.8932886","DOIUrl":null,"url":null,"abstract":"Camera-based pose estimation is a necessity for flexible applications in robotics, especially interaction between robots and mobile entities. Inspired by recent advancements in human pose estimation based on Convolutional Neural Networks, it is aspired to substitute the usage of artificial marker by automatically detecting inherent keypoints of the robot representing its 2D skeleton model. In addition, current encoder readings of the robot are utilized establishing the corresponding 3D skeleton model through forward kinematics. With the help of these 2D - 3D point correspondences, an estimation of the translation and orientation deviation between robot and camera is derived solving the perspective-n-point problem. An adequate approach for markerless keypoint detection of an UR5 robot is presented and evaluated in terms of precision and pose dispersion considering a dynamically moving robot. The promising results show that the novel method works robustly and reliably as a few-shot approach and copes with false positives as well as with partly occlusions and non-detected keypoints. Further potential is identified regarding enhancing the accuracy through the use of synthetic data.","PeriodicalId":234023,"journal":{"name":"2019 7th International Conference on Robot Intelligence Technology and Applications (RiTA)","volume":"46 2","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"13","resultStr":"{\"title\":\"Robust Few-Shot Pose Estimation of Articulated Robots using Monocular Cameras and Deep-Learning-based Keypoint Detection\",\"authors\":\"Jens Lambrecht\",\"doi\":\"10.1109/RITAPP.2019.8932886\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Camera-based pose estimation is a necessity for flexible applications in robotics, especially interaction between robots and mobile entities. Inspired by recent advancements in human pose estimation based on Convolutional Neural Networks, it is aspired to substitute the usage of artificial marker by automatically detecting inherent keypoints of the robot representing its 2D skeleton model. In addition, current encoder readings of the robot are utilized establishing the corresponding 3D skeleton model through forward kinematics. With the help of these 2D - 3D point correspondences, an estimation of the translation and orientation deviation between robot and camera is derived solving the perspective-n-point problem. An adequate approach for markerless keypoint detection of an UR5 robot is presented and evaluated in terms of precision and pose dispersion considering a dynamically moving robot. The promising results show that the novel method works robustly and reliably as a few-shot approach and copes with false positives as well as with partly occlusions and non-detected keypoints. Further potential is identified regarding enhancing the accuracy through the use of synthetic data.\",\"PeriodicalId\":234023,\"journal\":{\"name\":\"2019 7th International Conference on Robot Intelligence Technology and Applications (RiTA)\",\"volume\":\"46 2\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"13\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 7th International Conference on Robot Intelligence Technology and Applications (RiTA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/RITAPP.2019.8932886\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 7th International Conference on Robot Intelligence Technology and Applications (RiTA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/RITAPP.2019.8932886","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Robust Few-Shot Pose Estimation of Articulated Robots using Monocular Cameras and Deep-Learning-based Keypoint Detection
Camera-based pose estimation is a necessity for flexible applications in robotics, especially interaction between robots and mobile entities. Inspired by recent advancements in human pose estimation based on Convolutional Neural Networks, it is aspired to substitute the usage of artificial marker by automatically detecting inherent keypoints of the robot representing its 2D skeleton model. In addition, current encoder readings of the robot are utilized establishing the corresponding 3D skeleton model through forward kinematics. With the help of these 2D - 3D point correspondences, an estimation of the translation and orientation deviation between robot and camera is derived solving the perspective-n-point problem. An adequate approach for markerless keypoint detection of an UR5 robot is presented and evaluated in terms of precision and pose dispersion considering a dynamically moving robot. The promising results show that the novel method works robustly and reliably as a few-shot approach and copes with false positives as well as with partly occlusions and non-detected keypoints. Further potential is identified regarding enhancing the accuracy through the use of synthetic data.