Human-robot collaborative assembly task planning for mobile cobots based on deep reinforcement learning

IF 1.8 4区 工程技术 Q3 ENGINEERING, MECHANICAL
Wenbin Hou, Zhihua Xiong, Ming Yue, Hao Chen
{"title":"Human-robot collaborative assembly task planning for mobile cobots based on deep reinforcement learning","authors":"Wenbin Hou, Zhihua Xiong, Ming Yue, Hao Chen","doi":"10.1177/09544062241271626","DOIUrl":null,"url":null,"abstract":"Human-Robot Collaborative Assembly (HRCA) offers a novel solution to improve manual-based assembly manufacturing processes. However, the existing HRCA task planning approaches are limited in several aspects, such as inability to be applied effectively to mobile cobots with diverse workspaces and varying skill sets, as well as struggle in meeting the requirements of complex task structures and high-dimensional state spaces. In this paper, an interactive HRCA task planning framework based on finite Markov Decision Process (MDP) is presented, formalizing the Reinforcement Learning (RL) problem. To leverage the mobile and operational advantages of mobile cobots, a Multi-Attribute Hierarchical Task Network (MA-HTN) that enables efficient task decomposition and attribute representation is introduced. Additionally, considering state changes during assembly processes and utilizing Deep Reinforcement Learning (DRL) for handling high-dimensional decision problems, a universal DRL solving environment executed within unit time is constructed. This solving environment is based on a four-channel state diagram capable of reflecting high-dimensional state information that can be directly converted into digital tensor input for neural networks. Furthermore, to address frequent episode restarts in Deep Q-Network (DQN) algorithm and optimize task completion duration, a revival mechanism along with its enhanced algorithm are proposed. Finally, through an automobile fender bracket assembly scenario and an additional case study, the effectiveness of proposed method under varying numbers of tasks and work units is validated.","PeriodicalId":20558,"journal":{"name":"Proceedings of the Institution of Mechanical Engineers, Part C: Journal of Mechanical Engineering Science","volume":null,"pages":null},"PeriodicalIF":1.8000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the Institution of Mechanical Engineers, Part C: Journal of Mechanical Engineering Science","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1177/09544062241271626","RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, MECHANICAL","Score":null,"Total":0}
引用次数: 0

Abstract

Human-Robot Collaborative Assembly (HRCA) offers a novel solution to improve manual-based assembly manufacturing processes. However, the existing HRCA task planning approaches are limited in several aspects, such as inability to be applied effectively to mobile cobots with diverse workspaces and varying skill sets, as well as struggle in meeting the requirements of complex task structures and high-dimensional state spaces. In this paper, an interactive HRCA task planning framework based on finite Markov Decision Process (MDP) is presented, formalizing the Reinforcement Learning (RL) problem. To leverage the mobile and operational advantages of mobile cobots, a Multi-Attribute Hierarchical Task Network (MA-HTN) that enables efficient task decomposition and attribute representation is introduced. Additionally, considering state changes during assembly processes and utilizing Deep Reinforcement Learning (DRL) for handling high-dimensional decision problems, a universal DRL solving environment executed within unit time is constructed. This solving environment is based on a four-channel state diagram capable of reflecting high-dimensional state information that can be directly converted into digital tensor input for neural networks. Furthermore, to address frequent episode restarts in Deep Q-Network (DQN) algorithm and optimize task completion duration, a revival mechanism along with its enhanced algorithm are proposed. Finally, through an automobile fender bracket assembly scenario and an additional case study, the effectiveness of proposed method under varying numbers of tasks and work units is validated.
基于深度强化学习的移动协作机器人的人机协作装配任务规划
人机协作装配(Human-Robot Collaborative Assembly,HRCA)为改进基于人工的装配制造流程提供了一种新颖的解决方案。然而,现有的人机协作装配任务规划方法在多个方面存在局限性,例如无法有效应用于具有不同工作空间和不同技能组合的移动协作机器人,以及难以满足复杂任务结构和高维状态空间的要求。本文提出了一个基于有限马尔可夫决策过程(MDP)的交互式 HRCA 任务规划框架,将强化学习(RL)问题形式化。为了充分利用移动协作机器人的移动和操作优势,本文引入了多属性分层任务网络(MA-HTN),该网络可实现高效的任务分解和属性表示。此外,考虑到装配过程中的状态变化,并利用深度强化学习(DRL)处理高维决策问题,构建了在单位时间内执行的通用 DRL 求解环境。该求解环境基于四通道状态图,能够反映高维状态信息,并可直接转换为神经网络的数字张量输入。此外,为了解决深度 Q 网络(DQN)算法中频繁的情节重启问题并优化任务完成时间,提出了复兴机制及其增强算法。最后,通过汽车挡泥板支架装配场景和附加案例研究,验证了所提方法在不同任务数量和工作单元下的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
3.80
自引率
10.00%
发文量
625
审稿时长
4.3 months
期刊介绍: The Journal of Mechanical Engineering Science advances the understanding of both the fundamentals of engineering science and its application to the solution of challenges and problems in engineering.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信