Visual Imitation Learning for robotic fresh mushroom harvesting

Antonios Porichis, Konstantinos Vasios, Myrto Iglezou, Vishwanathan Mohan, P. Chatzakos
{"title":"Visual Imitation Learning for robotic fresh mushroom harvesting","authors":"Antonios Porichis, Konstantinos Vasios, Myrto Iglezou, Vishwanathan Mohan, P. Chatzakos","doi":"10.1109/MED59994.2023.10185745","DOIUrl":null,"url":null,"abstract":"Imitation Learning holds significant promise in enabling the automation of complex robotic manipulations tasks which are impossible to explicitly program. Mushroom harvesting is a task of high difficulty requiring weeks of intense training even for humans to master. In this work we present an end-to-end Imitation Learning pipeline that learns to apply the series of motions, namely reaching, grasping, twisting, and pulling the mushroom directly from pixel-level information. Mushroom harvesting experiments are carried out within a simulated environment that models the mushroom dynamics based on von Mises yielding theory with parameters obtained through expert picker demonstration wearing gloves with force sensors. We test the robustness of our technique by performing randomization on the camera extrinsic and intrinsic parameters as well as on the mushroom sizes. We also evaluate on different kinds of visual input namely grayscale and depth maps. Overall, our technique shows significant promise in automating mushroom harvesting directly from visual input while being remarkably lean in terms of computation intensity. Our models can be trained on a standard Laptop GPU in under one hour while inference of an action takes less than 1.5ms on a Laptop CPU. A brief overview of our experiments in video format is available at: https://bit.ly/41kCH7T","PeriodicalId":270226,"journal":{"name":"2023 31st Mediterranean Conference on Control and Automation (MED)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 31st Mediterranean Conference on Control and Automation (MED)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MED59994.2023.10185745","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Imitation Learning holds significant promise in enabling the automation of complex robotic manipulations tasks which are impossible to explicitly program. Mushroom harvesting is a task of high difficulty requiring weeks of intense training even for humans to master. In this work we present an end-to-end Imitation Learning pipeline that learns to apply the series of motions, namely reaching, grasping, twisting, and pulling the mushroom directly from pixel-level information. Mushroom harvesting experiments are carried out within a simulated environment that models the mushroom dynamics based on von Mises yielding theory with parameters obtained through expert picker demonstration wearing gloves with force sensors. We test the robustness of our technique by performing randomization on the camera extrinsic and intrinsic parameters as well as on the mushroom sizes. We also evaluate on different kinds of visual input namely grayscale and depth maps. Overall, our technique shows significant promise in automating mushroom harvesting directly from visual input while being remarkably lean in terms of computation intensity. Our models can be trained on a standard Laptop GPU in under one hour while inference of an action takes less than 1.5ms on a Laptop CPU. A brief overview of our experiments in video format is available at: https://bit.ly/41kCH7T
机器人鲜蘑菇收获的视觉模仿学习
模仿学习在使不可能明确编程的复杂机器人操作任务自动化方面具有重要的前景。蘑菇收获是一项高难度的任务,即使是人类也需要数周的高强度训练才能掌握。在这项工作中,我们提出了一个端到端的模仿学习管道,该管道学习应用一系列运动,即直接从像素级信息中到达,抓取,扭转和拉动蘑菇。在一个模拟环境中进行蘑菇采收实验,该环境基于von Mises屈服理论对蘑菇动力学进行建模,并通过戴有力传感器的手套的专家采摘演示获得参数。我们通过对相机的外在和内在参数以及蘑菇大小进行随机化来测试我们技术的稳健性。我们还评估了不同类型的视觉输入,即灰度图和深度图。总的来说,我们的技术在直接从视觉输入自动化蘑菇收获方面显示出巨大的希望,同时在计算强度方面非常精益。我们的模型可以在一个标准的笔记本电脑GPU上训练不到一个小时,而一个动作的推理在笔记本电脑CPU上需要不到1.5毫秒。我们的实验视频格式的简要概述可在:https://bit.ly/41kCH7T
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信