Hirohisa Kato, F. Nagata, Yuito Murakami, Keita Koya
{"title":"Partial Depth Estimation with Single Image Using YOLO and CNN for Robot Arm Control","authors":"Hirohisa Kato, F. Nagata, Yuito Murakami, Keita Koya","doi":"10.1109/ICMA54519.2022.9856055","DOIUrl":null,"url":null,"abstract":"This paper presents experimental results of partial object detection using YOLO (You Only Look Once) and partial depth estimation using CNN (Convolutional Neural Network) for application to a robot arm control. In recent years, image recognition and its application automation, as an alternative to advanced work, are attracting attention in various fields. In order for robot arms used in factories to perform high-value-added and flexible work, it is necessary to control them by object detection and recognition by deep learning. In this study, the authors propose a new approach for estimating the depth of partial images using YOLO and uses it to control the robot arm. In the experiments, both ends of a pen detected by YOLO are used for the input to a CNN. The detected parts are saved to images with a size of about 60 × 60 pixels, and the depths are estimated by giving the cropped images to the CNN. A desktop-sized robot with 4DOFs can successfully pick the pen by referring the depths. The effectiveness of the proposed method is demonstrated through experiments.","PeriodicalId":120073,"journal":{"name":"2022 IEEE International Conference on Mechatronics and Automation (ICMA)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE International Conference on Mechatronics and Automation (ICMA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMA54519.2022.9856055","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
This paper presents experimental results of partial object detection using YOLO (You Only Look Once) and partial depth estimation using CNN (Convolutional Neural Network) for application to a robot arm control. In recent years, image recognition and its application automation, as an alternative to advanced work, are attracting attention in various fields. In order for robot arms used in factories to perform high-value-added and flexible work, it is necessary to control them by object detection and recognition by deep learning. In this study, the authors propose a new approach for estimating the depth of partial images using YOLO and uses it to control the robot arm. In the experiments, both ends of a pen detected by YOLO are used for the input to a CNN. The detected parts are saved to images with a size of about 60 × 60 pixels, and the depths are estimated by giving the cropped images to the CNN. A desktop-sized robot with 4DOFs can successfully pick the pen by referring the depths. The effectiveness of the proposed method is demonstrated through experiments.