{"title":"A vision-language-guided and deep reinforcement learning-enabled approach for unstructured human-robot collaborative manufacturing task fulfilment","authors":"","doi":"10.1016/j.cirp.2024.04.003","DOIUrl":null,"url":null,"abstract":"<div><p>Human-Robot Collaboration (HRC) has emerged as a pivot in contemporary human-centric smart manufacturing scenarios. However, the fulfilment of HRC tasks in unstructured scenes brings many challenges to be overcome. In this work, mixed reality head-mounted display is modelled as an effective data collection, communication, and state representation interface/tool for HRC task settings. By integrating vision-language cues with large language model, a vision-language-guided HRC task planning approach is firstly proposed. Then, a deep reinforcement learning-enabled mobile manipulator motion control policy is generated to fulfil HRC task primitives. Its feasibility is demonstrated in several HRC unstructured manufacturing tasks with comparative results.</p></div>","PeriodicalId":55256,"journal":{"name":"Cirp Annals-Manufacturing Technology","volume":"73 1","pages":"Pages 341-344"},"PeriodicalIF":3.2000,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0007850624000180/pdfft?md5=d49fa8d97038e77f13844355ec91f7eb&pid=1-s2.0-S0007850624000180-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Cirp Annals-Manufacturing Technology","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0007850624000180","RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, INDUSTRIAL","Score":null,"Total":0}
引用次数: 0
Abstract
Human-Robot Collaboration (HRC) has emerged as a pivot in contemporary human-centric smart manufacturing scenarios. However, the fulfilment of HRC tasks in unstructured scenes brings many challenges to be overcome. In this work, mixed reality head-mounted display is modelled as an effective data collection, communication, and state representation interface/tool for HRC task settings. By integrating vision-language cues with large language model, a vision-language-guided HRC task planning approach is firstly proposed. Then, a deep reinforcement learning-enabled mobile manipulator motion control policy is generated to fulfil HRC task primitives. Its feasibility is demonstrated in several HRC unstructured manufacturing tasks with comparative results.
期刊介绍:
CIRP, The International Academy for Production Engineering, was founded in 1951 to promote, by scientific research, the development of all aspects of manufacturing technology covering the optimization, control and management of processes, machines and systems.
This biannual ISI cited journal contains approximately 140 refereed technical and keynote papers. Subject areas covered include:
Assembly, Cutting, Design, Electro-Physical and Chemical Processes, Forming, Abrasive processes, Surfaces, Machines, Production Systems and Organizations, Precision Engineering and Metrology, Life-Cycle Engineering, Microsystems Technology (MST), Nanotechnology.