Programming by Visual Demonstration for Pick-and-Place Tasks using Robot Skills

Peng Hao, Tao Lu, Yinghao Cai, Shuo Wang
{"title":"Programming by Visual Demonstration for Pick-and-Place Tasks using Robot Skills","authors":"Peng Hao, Tao Lu, Yinghao Cai, Shuo Wang","doi":"10.1109/ROBIO49542.2019.8961481","DOIUrl":null,"url":null,"abstract":"In this paper, we present a vision-based robot programming system for pick-and-place tasks that can generate programs from human demonstrations. The system consists of a detection network and a program generation module. The detection network leverages convolutional pose machines to detect the key-points of the objects. The network is trained in a simulation environment in which the train set is collected and auto-labeled. To bridge the gap between reality and simulation, we propose a design method of transform function for mapping a real image to synthesized style. Compared with the unmapped results, the Mean Absolute Error (MAE) of the model completely trained with synthesized images is reduced by 23% and the False Negative Rate FNR (FNR) of the model fine-tuned by the real images is reduced by 42.5% after mapping. The program generation module provides a human-readable program based on the detection results to reproduce a real-world demonstration, in which a longshort memory (LSM) is designed to integrate current and historical information. The system is tested in the real world with a UR5 robot on the task of stacking colored cubes in different orders.","PeriodicalId":121822,"journal":{"name":"2019 IEEE International Conference on Robotics and Biomimetics (ROBIO)","volume":"67 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE International Conference on Robotics and Biomimetics (ROBIO)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ROBIO49542.2019.8961481","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

In this paper, we present a vision-based robot programming system for pick-and-place tasks that can generate programs from human demonstrations. The system consists of a detection network and a program generation module. The detection network leverages convolutional pose machines to detect the key-points of the objects. The network is trained in a simulation environment in which the train set is collected and auto-labeled. To bridge the gap between reality and simulation, we propose a design method of transform function for mapping a real image to synthesized style. Compared with the unmapped results, the Mean Absolute Error (MAE) of the model completely trained with synthesized images is reduced by 23% and the False Negative Rate FNR (FNR) of the model fine-tuned by the real images is reduced by 42.5% after mapping. The program generation module provides a human-readable program based on the detection results to reproduce a real-world demonstration, in which a longshort memory (LSM) is designed to integrate current and historical information. The system is tested in the real world with a UR5 robot on the task of stacking colored cubes in different orders.
使用机器人技能的拾取任务的可视化演示编程
在本文中,我们提出了一个基于视觉的机器人编程系统,用于拾取和放置任务,该系统可以从人类演示中生成程序。该系统由检测网络和程序生成模块组成。检测网络利用卷积姿态机来检测物体的关键点。该网络在模拟环境中进行训练,在模拟环境中收集训练集并自动标记。为了弥合现实与仿真之间的差距,我们提出了一种将真实图像映射到合成样式的变换函数设计方法。与未映射的结果相比,完全使用合成图像训练的模型的平均绝对误差(MAE)降低了23%,经真实图像微调后的模型的假阴性率(FNR)降低了42.5%。程序生成模块根据检测结果提供人类可读的程序来再现真实世界的演示,其中设计了长短存储器(LSM)来集成当前和历史信息。该系统在现实世界中用UR5机器人测试了以不同顺序堆叠彩色立方体的任务。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信