A parallel graph network for generating 7-DoF model-free grasps in unstructured scenes using point cloud

IF 9.1 1区 计算机科学 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS
Chungang Zhuang, Haowen Wang, Wanhao Niu, Han Ding
{"title":"A parallel graph network for generating 7-DoF model-free grasps in unstructured scenes using point cloud","authors":"Chungang Zhuang,&nbsp;Haowen Wang,&nbsp;Wanhao Niu,&nbsp;Han Ding","doi":"10.1016/j.rcim.2024.102879","DOIUrl":null,"url":null,"abstract":"<div><p>Generating model-free grasps in complex scattered scenes remains a challenging task. Most current methods adopt PointNet++ as the backbone to extract structural features, while the relative associations of geometry are underexplored, leading to non-optimal grasp prediction results. In this work, a parallelized graph-based pipeline is developed to solve the 7-DoF grasp pose generation problem with point cloud as input. Using the non-textured information of the grasping scene, the proposed pipeline simultaneously performs feature embedding and grasping location focusing in two branches, avoiding the mutual influence of the two learning processes. In the feature learning branch, the geometric features of the whole scene will be fully learned. In the location focusing branch, the high-value grasping locations on the surface of objects will be strategically selected. Using the learned graph features at these locations, the pipeline will eventually output refined grasping directions and widths in conjunction with local spatial features. To strengthen the positional features in the grasping problem, a graph convolution operator based on the positional attention mechanism is designed, and a graph residual network based on this operator is applied in two branches. The above pipeline abstracts the grasping location selection task from the main process of grasp generation, which lowers the learning difficulty while avoiding the performance degradation problem of deep graph networks. The established pipeline is evaluated on the GraspNet-1Billion dataset, demonstrating much better performance and stronger generalization capabilities than the benchmark approach. In robotic bin-picking experiments, the proposed method can effectively understand scattered grasping scenarios and grasp multiple types of unknown objects with a high success rate.</p></div>","PeriodicalId":21452,"journal":{"name":"Robotics and Computer-integrated Manufacturing","volume":"92 ","pages":"Article 102879"},"PeriodicalIF":9.1000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Robotics and Computer-integrated Manufacturing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0736584524001662","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
引用次数: 0

Abstract

Generating model-free grasps in complex scattered scenes remains a challenging task. Most current methods adopt PointNet++ as the backbone to extract structural features, while the relative associations of geometry are underexplored, leading to non-optimal grasp prediction results. In this work, a parallelized graph-based pipeline is developed to solve the 7-DoF grasp pose generation problem with point cloud as input. Using the non-textured information of the grasping scene, the proposed pipeline simultaneously performs feature embedding and grasping location focusing in two branches, avoiding the mutual influence of the two learning processes. In the feature learning branch, the geometric features of the whole scene will be fully learned. In the location focusing branch, the high-value grasping locations on the surface of objects will be strategically selected. Using the learned graph features at these locations, the pipeline will eventually output refined grasping directions and widths in conjunction with local spatial features. To strengthen the positional features in the grasping problem, a graph convolution operator based on the positional attention mechanism is designed, and a graph residual network based on this operator is applied in two branches. The above pipeline abstracts the grasping location selection task from the main process of grasp generation, which lowers the learning difficulty while avoiding the performance degradation problem of deep graph networks. The established pipeline is evaluated on the GraspNet-1Billion dataset, demonstrating much better performance and stronger generalization capabilities than the benchmark approach. In robotic bin-picking experiments, the proposed method can effectively understand scattered grasping scenarios and grasp multiple types of unknown objects with a high success rate.

利用点云在非结构化场景中生成 7-DoF 无模型抓手的并行图网络
在复杂的零散场景中生成无模型抓取仍然是一项具有挑战性的任务。目前的大多数方法都采用 PointNet++ 作为提取结构特征的骨干,而对几何图形的相对关联探索不足,导致抓取预测结果不理想。在这项工作中,开发了一种基于图的并行化流水线,用于解决以点云为输入的 7-DoF 抓姿生成问题。利用抓取场景的非纹理信息,所提出的流水线在两个分支中同时执行特征嵌入和抓取位置聚焦,避免了两个学习过程的相互影响。在特征学习分支中,整个场景的几何特征将被完全学习。在位置聚焦分支中,将战略性地选择物体表面的高价值抓取位置。利用在这些位置学习到的图形特征,管道将结合局部空间特征,最终输出细化的抓取方向和宽度。为了强化抓取问题中的位置特征,我们设计了基于位置注意力机制的图卷积算子,并在两个分支中应用了基于该算子的图残差网络。上述管道将抓取位置选择任务从抓取生成的主要过程中抽象出来,降低了学习难度,同时避免了深度图网络的性能下降问题。在 GraspNet-1Billion 数据集上对所建立的管道进行了评估,结果表明其性能和泛化能力远远优于基准方法。在机器人分拣实验中,所提出的方法能有效地理解分散的抓取场景,并以较高的成功率抓取多种类型的未知物体。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Robotics and Computer-integrated Manufacturing
Robotics and Computer-integrated Manufacturing 工程技术-工程:制造
CiteScore
24.10
自引率
13.50%
发文量
160
审稿时长
50 days
期刊介绍: The journal, Robotics and Computer-Integrated Manufacturing, focuses on sharing research applications that contribute to the development of new or enhanced robotics, manufacturing technologies, and innovative manufacturing strategies that are relevant to industry. Papers that combine theory and experimental validation are preferred, while review papers on current robotics and manufacturing issues are also considered. However, papers on traditional machining processes, modeling and simulation, supply chain management, and resource optimization are generally not within the scope of the journal, as there are more appropriate journals for these topics. Similarly, papers that are overly theoretical or mathematical will be directed to other suitable journals. The journal welcomes original papers in areas such as industrial robotics, human-robot collaboration in manufacturing, cloud-based manufacturing, cyber-physical production systems, big data analytics in manufacturing, smart mechatronics, machine learning, adaptive and sustainable manufacturing, and other fields involving unique manufacturing technologies.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信