{"title":"基于严格训练卷积神经网络的实时目标坐标检测与机械手控制","authors":"Yu-Ming Chang, C. G. Li, Yi-Feng Hong","doi":"10.1109/COASE.2019.8842973","DOIUrl":null,"url":null,"abstract":"Objects embedded in the environment, such as switches, control buttons, sockets, et al., are devices that need frequent operations. To devise manipulators to operate such devices automatically, we propose a visual-position control scheme that directly converts the visual coordinate detections to motor commands. We train ConvNets with rigid 3D coordinate information, which is obtained from a single basis image of the target object. Our proposed training data preparation frameworks automatically generate and organize the required structure of the training images for the network. The ConvNet’s superior image recognition capability results in high success rate in object detection and high precision in coordinate estimation. In our static experiments, in-range plane coordinate detection achieves an average success rate of 91% from various view-point directions; the depth coordinate detection achieves an average success rate of 86% based on an extended success range. In our dynamic experiments, a low-precision manipulator was used to press a down elevator call button and achieved an overall success rate of 98%. A high-precision manipulator was used for an object localization task and achieved a precision of ± 0.3 mm using a low-resolution camera.","PeriodicalId":6695,"journal":{"name":"2019 IEEE 15th International Conference on Automation Science and Engineering (CASE)","volume":"76 1","pages":"1347-1352"},"PeriodicalIF":0.0000,"publicationDate":"2019-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":"{\"title\":\"Real-Time Object Coordinate Detection and Manipulator Control Using Rigidly Trained Convolutional Neural Networks\",\"authors\":\"Yu-Ming Chang, C. G. Li, Yi-Feng Hong\",\"doi\":\"10.1109/COASE.2019.8842973\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Objects embedded in the environment, such as switches, control buttons, sockets, et al., are devices that need frequent operations. To devise manipulators to operate such devices automatically, we propose a visual-position control scheme that directly converts the visual coordinate detections to motor commands. We train ConvNets with rigid 3D coordinate information, which is obtained from a single basis image of the target object. Our proposed training data preparation frameworks automatically generate and organize the required structure of the training images for the network. The ConvNet’s superior image recognition capability results in high success rate in object detection and high precision in coordinate estimation. In our static experiments, in-range plane coordinate detection achieves an average success rate of 91% from various view-point directions; the depth coordinate detection achieves an average success rate of 86% based on an extended success range. In our dynamic experiments, a low-precision manipulator was used to press a down elevator call button and achieved an overall success rate of 98%. A high-precision manipulator was used for an object localization task and achieved a precision of ± 0.3 mm using a low-resolution camera.\",\"PeriodicalId\":6695,\"journal\":{\"name\":\"2019 IEEE 15th International Conference on Automation Science and Engineering (CASE)\",\"volume\":\"76 1\",\"pages\":\"1347-1352\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"6\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 IEEE 15th International Conference on Automation Science and Engineering (CASE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/COASE.2019.8842973\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE 15th International Conference on Automation Science and Engineering (CASE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/COASE.2019.8842973","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Real-Time Object Coordinate Detection and Manipulator Control Using Rigidly Trained Convolutional Neural Networks
Objects embedded in the environment, such as switches, control buttons, sockets, et al., are devices that need frequent operations. To devise manipulators to operate such devices automatically, we propose a visual-position control scheme that directly converts the visual coordinate detections to motor commands. We train ConvNets with rigid 3D coordinate information, which is obtained from a single basis image of the target object. Our proposed training data preparation frameworks automatically generate and organize the required structure of the training images for the network. The ConvNet’s superior image recognition capability results in high success rate in object detection and high precision in coordinate estimation. In our static experiments, in-range plane coordinate detection achieves an average success rate of 91% from various view-point directions; the depth coordinate detection achieves an average success rate of 86% based on an extended success range. In our dynamic experiments, a low-precision manipulator was used to press a down elevator call button and achieved an overall success rate of 98%. A high-precision manipulator was used for an object localization task and achieved a precision of ± 0.3 mm using a low-resolution camera.