Antonios Porichis, Konstantinos Vasios, Myrto Iglezou, Vishwanathan Mohan, P. Chatzakos
{"title":"机器人鲜蘑菇收获的视觉模仿学习","authors":"Antonios Porichis, Konstantinos Vasios, Myrto Iglezou, Vishwanathan Mohan, P. Chatzakos","doi":"10.1109/MED59994.2023.10185745","DOIUrl":null,"url":null,"abstract":"Imitation Learning holds significant promise in enabling the automation of complex robotic manipulations tasks which are impossible to explicitly program. Mushroom harvesting is a task of high difficulty requiring weeks of intense training even for humans to master. In this work we present an end-to-end Imitation Learning pipeline that learns to apply the series of motions, namely reaching, grasping, twisting, and pulling the mushroom directly from pixel-level information. Mushroom harvesting experiments are carried out within a simulated environment that models the mushroom dynamics based on von Mises yielding theory with parameters obtained through expert picker demonstration wearing gloves with force sensors. We test the robustness of our technique by performing randomization on the camera extrinsic and intrinsic parameters as well as on the mushroom sizes. We also evaluate on different kinds of visual input namely grayscale and depth maps. Overall, our technique shows significant promise in automating mushroom harvesting directly from visual input while being remarkably lean in terms of computation intensity. Our models can be trained on a standard Laptop GPU in under one hour while inference of an action takes less than 1.5ms on a Laptop CPU. A brief overview of our experiments in video format is available at: https://bit.ly/41kCH7T","PeriodicalId":270226,"journal":{"name":"2023 31st Mediterranean Conference on Control and Automation (MED)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Visual Imitation Learning for robotic fresh mushroom harvesting\",\"authors\":\"Antonios Porichis, Konstantinos Vasios, Myrto Iglezou, Vishwanathan Mohan, P. Chatzakos\",\"doi\":\"10.1109/MED59994.2023.10185745\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Imitation Learning holds significant promise in enabling the automation of complex robotic manipulations tasks which are impossible to explicitly program. Mushroom harvesting is a task of high difficulty requiring weeks of intense training even for humans to master. In this work we present an end-to-end Imitation Learning pipeline that learns to apply the series of motions, namely reaching, grasping, twisting, and pulling the mushroom directly from pixel-level information. Mushroom harvesting experiments are carried out within a simulated environment that models the mushroom dynamics based on von Mises yielding theory with parameters obtained through expert picker demonstration wearing gloves with force sensors. We test the robustness of our technique by performing randomization on the camera extrinsic and intrinsic parameters as well as on the mushroom sizes. We also evaluate on different kinds of visual input namely grayscale and depth maps. Overall, our technique shows significant promise in automating mushroom harvesting directly from visual input while being remarkably lean in terms of computation intensity. Our models can be trained on a standard Laptop GPU in under one hour while inference of an action takes less than 1.5ms on a Laptop CPU. A brief overview of our experiments in video format is available at: https://bit.ly/41kCH7T\",\"PeriodicalId\":270226,\"journal\":{\"name\":\"2023 31st Mediterranean Conference on Control and Automation (MED)\",\"volume\":\"6 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-06-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 31st Mediterranean Conference on Control and Automation (MED)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/MED59994.2023.10185745\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 31st Mediterranean Conference on Control and Automation (MED)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MED59994.2023.10185745","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Visual Imitation Learning for robotic fresh mushroom harvesting
Imitation Learning holds significant promise in enabling the automation of complex robotic manipulations tasks which are impossible to explicitly program. Mushroom harvesting is a task of high difficulty requiring weeks of intense training even for humans to master. In this work we present an end-to-end Imitation Learning pipeline that learns to apply the series of motions, namely reaching, grasping, twisting, and pulling the mushroom directly from pixel-level information. Mushroom harvesting experiments are carried out within a simulated environment that models the mushroom dynamics based on von Mises yielding theory with parameters obtained through expert picker demonstration wearing gloves with force sensors. We test the robustness of our technique by performing randomization on the camera extrinsic and intrinsic parameters as well as on the mushroom sizes. We also evaluate on different kinds of visual input namely grayscale and depth maps. Overall, our technique shows significant promise in automating mushroom harvesting directly from visual input while being remarkably lean in terms of computation intensity. Our models can be trained on a standard Laptop GPU in under one hour while inference of an action takes less than 1.5ms on a Laptop CPU. A brief overview of our experiments in video format is available at: https://bit.ly/41kCH7T