{"title":"非结构化人机协作制造任务完成的视觉语言引导和深度强化学习方法","authors":"","doi":"10.1016/j.cirp.2024.04.003","DOIUrl":null,"url":null,"abstract":"<div><p>Human-Robot Collaboration (HRC) has emerged as a pivot in contemporary human-centric smart manufacturing scenarios. However, the fulfilment of HRC tasks in unstructured scenes brings many challenges to be overcome. In this work, mixed reality head-mounted display is modelled as an effective data collection, communication, and state representation interface/tool for HRC task settings. By integrating vision-language cues with large language model, a vision-language-guided HRC task planning approach is firstly proposed. Then, a deep reinforcement learning-enabled mobile manipulator motion control policy is generated to fulfil HRC task primitives. Its feasibility is demonstrated in several HRC unstructured manufacturing tasks with comparative results.</p></div>","PeriodicalId":55256,"journal":{"name":"Cirp Annals-Manufacturing Technology","volume":"73 1","pages":"Pages 341-344"},"PeriodicalIF":3.2000,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0007850624000180/pdfft?md5=d49fa8d97038e77f13844355ec91f7eb&pid=1-s2.0-S0007850624000180-main.pdf","citationCount":"0","resultStr":"{\"title\":\"A vision-language-guided and deep reinforcement learning-enabled approach for unstructured human-robot collaborative manufacturing task fulfilment\",\"authors\":\"\",\"doi\":\"10.1016/j.cirp.2024.04.003\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Human-Robot Collaboration (HRC) has emerged as a pivot in contemporary human-centric smart manufacturing scenarios. However, the fulfilment of HRC tasks in unstructured scenes brings many challenges to be overcome. In this work, mixed reality head-mounted display is modelled as an effective data collection, communication, and state representation interface/tool for HRC task settings. By integrating vision-language cues with large language model, a vision-language-guided HRC task planning approach is firstly proposed. Then, a deep reinforcement learning-enabled mobile manipulator motion control policy is generated to fulfil HRC task primitives. Its feasibility is demonstrated in several HRC unstructured manufacturing tasks with comparative results.</p></div>\",\"PeriodicalId\":55256,\"journal\":{\"name\":\"Cirp Annals-Manufacturing Technology\",\"volume\":\"73 1\",\"pages\":\"Pages 341-344\"},\"PeriodicalIF\":3.2000,\"publicationDate\":\"2024-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.sciencedirect.com/science/article/pii/S0007850624000180/pdfft?md5=d49fa8d97038e77f13844355ec91f7eb&pid=1-s2.0-S0007850624000180-main.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Cirp Annals-Manufacturing Technology\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0007850624000180\",\"RegionNum\":3,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ENGINEERING, INDUSTRIAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Cirp Annals-Manufacturing Technology","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0007850624000180","RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, INDUSTRIAL","Score":null,"Total":0}
A vision-language-guided and deep reinforcement learning-enabled approach for unstructured human-robot collaborative manufacturing task fulfilment
Human-Robot Collaboration (HRC) has emerged as a pivot in contemporary human-centric smart manufacturing scenarios. However, the fulfilment of HRC tasks in unstructured scenes brings many challenges to be overcome. In this work, mixed reality head-mounted display is modelled as an effective data collection, communication, and state representation interface/tool for HRC task settings. By integrating vision-language cues with large language model, a vision-language-guided HRC task planning approach is firstly proposed. Then, a deep reinforcement learning-enabled mobile manipulator motion control policy is generated to fulfil HRC task primitives. Its feasibility is demonstrated in several HRC unstructured manufacturing tasks with comparative results.
期刊介绍:
CIRP, The International Academy for Production Engineering, was founded in 1951 to promote, by scientific research, the development of all aspects of manufacturing technology covering the optimization, control and management of processes, machines and systems.
This biannual ISI cited journal contains approximately 140 refereed technical and keynote papers. Subject areas covered include:
Assembly, Cutting, Design, Electro-Physical and Chemical Processes, Forming, Abrasive processes, Surfaces, Machines, Production Systems and Organizations, Precision Engineering and Metrology, Life-Cycle Engineering, Microsystems Technology (MST), Nanotechnology.