Guoyu Zuo, Jiayuan Tong, Hongxing Liu, Wenbai Chen, Jianfeng Li
{"title":"Graph-based Visual Manipulation Relationship Reasoning in Object-Stacking Scenes","authors":"Guoyu Zuo, Jiayuan Tong, Hongxing Liu, Wenbai Chen, Jianfeng Li","doi":"10.1109/IJCNN52387.2021.9534389","DOIUrl":null,"url":null,"abstract":"In object-stacking scenes, robotic manipulation is one of the most important research topics in robotics. It is particularly significant to reason object relationships and obtain intelligent manipulation order for more advanced interaction between the robot and the environment. However, many existing methods focus on individual object features and ignore contextual information, leading to great challenges in efficiently reasoning manipulation relationship. In this paper, we introduce a novel graph-based visual manipulation relationship reasoning architecture that directly outputs object relationships and manipulation order. Our model first extracts features and detects objects from RGB images, and then adopts Graph Convolutional Network (GCN) to collect contextual information between objects. Moreover, a relationship filtering network is built to reduce object pairs before reasoning and improve the efficiency of relation reasoning. The experiments on the Visual Manipulation Relationship Dataset (VMRD) show that our model significantly outperforms previous methods on reasoning object relationships in obiect-stackina scenes.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Joint Conference on Neural Networks (IJCNN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IJCNN52387.2021.9534389","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
Abstract
In object-stacking scenes, robotic manipulation is one of the most important research topics in robotics. It is particularly significant to reason object relationships and obtain intelligent manipulation order for more advanced interaction between the robot and the environment. However, many existing methods focus on individual object features and ignore contextual information, leading to great challenges in efficiently reasoning manipulation relationship. In this paper, we introduce a novel graph-based visual manipulation relationship reasoning architecture that directly outputs object relationships and manipulation order. Our model first extracts features and detects objects from RGB images, and then adopts Graph Convolutional Network (GCN) to collect contextual information between objects. Moreover, a relationship filtering network is built to reduce object pairs before reasoning and improve the efficiency of relation reasoning. The experiments on the Visual Manipulation Relationship Dataset (VMRD) show that our model significantly outperforms previous methods on reasoning object relationships in obiect-stackina scenes.