{"title":"Robotic Grasp Detection by Rotation Region CNN","authors":"Hsien-I Lin, Hong-Qi Chu","doi":"10.1109/INDIN45523.2021.9557573","DOIUrl":null,"url":null,"abstract":"Recently using deep learning methods for robotic grasping is a promising research. Many previous works used one-or two-stage deep learning methods to learn optimal grasping rectangles. However, these deep learning methods mainly detected vertical bounding boxes and performed post-processing for finding grasps. To avoid post-processing, we adopt the rotation region convolutional neural network (R2CNN) to detect oriented optimal grasps without post-preprocess. The modified R2CNN is divided into three stages: (1) feature extraction, (2) intermediate layer, and (3) gasp detection. In the second stage, we found that using a smaller set of anchor scale and a small IoU threshold were helpful to detect correct grasping rectangles. In our experiment, we used the Cornell grasping dataset as the benchmark and validated that using both axis-aligned and inclined bounding boxes in training. The results show that our modified R2CNN for image-wise detection reached up to 96% in accuracy.","PeriodicalId":370921,"journal":{"name":"2021 IEEE 19th International Conference on Industrial Informatics (INDIN)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 19th International Conference on Industrial Informatics (INDIN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/INDIN45523.2021.9557573","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
Recently using deep learning methods for robotic grasping is a promising research. Many previous works used one-or two-stage deep learning methods to learn optimal grasping rectangles. However, these deep learning methods mainly detected vertical bounding boxes and performed post-processing for finding grasps. To avoid post-processing, we adopt the rotation region convolutional neural network (R2CNN) to detect oriented optimal grasps without post-preprocess. The modified R2CNN is divided into three stages: (1) feature extraction, (2) intermediate layer, and (3) gasp detection. In the second stage, we found that using a smaller set of anchor scale and a small IoU threshold were helpful to detect correct grasping rectangles. In our experiment, we used the Cornell grasping dataset as the benchmark and validated that using both axis-aligned and inclined bounding boxes in training. The results show that our modified R2CNN for image-wise detection reached up to 96% in accuracy.