Tat Hieu Bui;Yeong Gwang Son;Juyong Hong;Yong Hyeon Kim;Hyouk Ryeol Choi
{"title":"通过学习和分析融合增强复杂环境中的对映抓取模式","authors":"Tat Hieu Bui;Yeong Gwang Son;Juyong Hong;Yong Hyeon Kim;Hyouk Ryeol Choi","doi":"10.1109/TASE.2024.3512005","DOIUrl":null,"url":null,"abstract":"The robotic pick-and-place system is applied widely in many fields such as assembly, packaging, bin-picking, and sorting. In this paper, we present a deep learning and analytical-based method for generating antipodal grasping multi-modality in highly cluttered scenes. Our method takes advantage of three types of grasp poses to deal with the complexity of environment and achieves efficient computation time for real applications. A new synthetic training datasets are generated in Isaac Sim including approximately 35000 RGB-D images and an automatic labeling algorithm is developed. We utilize convolutional neural networks (CNNs) for predicting antipodal grasping parameters on objects and a filtering algorithm to avoid collisions and calculate grasp’s depth simultaneously. Our approach processes the entire task in approximately 0.2 seconds, achieving a success rate of over 96% and more than 98% collision-free grasps in cluttered scenes. The method was verified by experiments with RB10 robot arm, two-fingers grippers, depth camera L515, and several objects in different scenes. The article shows a simple, effective, and highly applicable approach in real environments. The real experimental video of our method is shown at <uri>https://www.youtube.com/watch?v=GvJZxUyQr3w</uri>. Note to Practitioners—The robotic pick-and-place task is fundamental to the progress of automation. This article presents a novel method for efficiently picking objects from complex environments characterized by obstacles, overlaps, and a diversity of objects. These scenarios pose challenges in approaching objects and avoiding collisions. Our method combines learning and analytical approaches, demonstrating high performance in terms of accuracy, generalization across various scenes and grippers, and rapid computation, with the entire process taking approximately 0.2 seconds. These advancements are validated through a series of real robotic grasping experiments and comparisons with state-of-the-art methods. We believe that our method represents a significant contribution to the field of automation science and holds promise for real-world applications.","PeriodicalId":51060,"journal":{"name":"IEEE Transactions on Automation Science and Engineering","volume":"22 ","pages":"9767-9781"},"PeriodicalIF":6.4000,"publicationDate":"2024-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Enhancing Antipodal Grasping Modalities in Complex Environments Through Learning and Analytical Fusion\",\"authors\":\"Tat Hieu Bui;Yeong Gwang Son;Juyong Hong;Yong Hyeon Kim;Hyouk Ryeol Choi\",\"doi\":\"10.1109/TASE.2024.3512005\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The robotic pick-and-place system is applied widely in many fields such as assembly, packaging, bin-picking, and sorting. In this paper, we present a deep learning and analytical-based method for generating antipodal grasping multi-modality in highly cluttered scenes. Our method takes advantage of three types of grasp poses to deal with the complexity of environment and achieves efficient computation time for real applications. A new synthetic training datasets are generated in Isaac Sim including approximately 35000 RGB-D images and an automatic labeling algorithm is developed. We utilize convolutional neural networks (CNNs) for predicting antipodal grasping parameters on objects and a filtering algorithm to avoid collisions and calculate grasp’s depth simultaneously. Our approach processes the entire task in approximately 0.2 seconds, achieving a success rate of over 96% and more than 98% collision-free grasps in cluttered scenes. The method was verified by experiments with RB10 robot arm, two-fingers grippers, depth camera L515, and several objects in different scenes. The article shows a simple, effective, and highly applicable approach in real environments. The real experimental video of our method is shown at <uri>https://www.youtube.com/watch?v=GvJZxUyQr3w</uri>. Note to Practitioners—The robotic pick-and-place task is fundamental to the progress of automation. This article presents a novel method for efficiently picking objects from complex environments characterized by obstacles, overlaps, and a diversity of objects. These scenarios pose challenges in approaching objects and avoiding collisions. Our method combines learning and analytical approaches, demonstrating high performance in terms of accuracy, generalization across various scenes and grippers, and rapid computation, with the entire process taking approximately 0.2 seconds. These advancements are validated through a series of real robotic grasping experiments and comparisons with state-of-the-art methods. We believe that our method represents a significant contribution to the field of automation science and holds promise for real-world applications.\",\"PeriodicalId\":51060,\"journal\":{\"name\":\"IEEE Transactions on Automation Science and Engineering\",\"volume\":\"22 \",\"pages\":\"9767-9781\"},\"PeriodicalIF\":6.4000,\"publicationDate\":\"2024-12-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Automation Science and Engineering\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10810737/\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Automation Science and Engineering","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10810737/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
Enhancing Antipodal Grasping Modalities in Complex Environments Through Learning and Analytical Fusion
The robotic pick-and-place system is applied widely in many fields such as assembly, packaging, bin-picking, and sorting. In this paper, we present a deep learning and analytical-based method for generating antipodal grasping multi-modality in highly cluttered scenes. Our method takes advantage of three types of grasp poses to deal with the complexity of environment and achieves efficient computation time for real applications. A new synthetic training datasets are generated in Isaac Sim including approximately 35000 RGB-D images and an automatic labeling algorithm is developed. We utilize convolutional neural networks (CNNs) for predicting antipodal grasping parameters on objects and a filtering algorithm to avoid collisions and calculate grasp’s depth simultaneously. Our approach processes the entire task in approximately 0.2 seconds, achieving a success rate of over 96% and more than 98% collision-free grasps in cluttered scenes. The method was verified by experiments with RB10 robot arm, two-fingers grippers, depth camera L515, and several objects in different scenes. The article shows a simple, effective, and highly applicable approach in real environments. The real experimental video of our method is shown at https://www.youtube.com/watch?v=GvJZxUyQr3w. Note to Practitioners—The robotic pick-and-place task is fundamental to the progress of automation. This article presents a novel method for efficiently picking objects from complex environments characterized by obstacles, overlaps, and a diversity of objects. These scenarios pose challenges in approaching objects and avoiding collisions. Our method combines learning and analytical approaches, demonstrating high performance in terms of accuracy, generalization across various scenes and grippers, and rapid computation, with the entire process taking approximately 0.2 seconds. These advancements are validated through a series of real robotic grasping experiments and comparisons with state-of-the-art methods. We believe that our method represents a significant contribution to the field of automation science and holds promise for real-world applications.
期刊介绍:
The IEEE Transactions on Automation Science and Engineering (T-ASE) publishes fundamental papers on Automation, emphasizing scientific results that advance efficiency, quality, productivity, and reliability. T-ASE encourages interdisciplinary approaches from computer science, control systems, electrical engineering, mathematics, mechanical engineering, operations research, and other fields. T-ASE welcomes results relevant to industries such as agriculture, biotechnology, healthcare, home automation, maintenance, manufacturing, pharmaceuticals, retail, security, service, supply chains, and transportation. T-ASE addresses a research community willing to integrate knowledge across disciplines and industries. For this purpose, each paper includes a Note to Practitioners that summarizes how its results can be applied or how they might be extended to apply in practice.