{"title":"Towards instructional collaborative robots: From video-based learning to feedback-adapted instruction","authors":"Jinyi Huang, Xun Xu, Jan Polzer","doi":"10.1016/j.jmsy.2025.08.018","DOIUrl":null,"url":null,"abstract":"<div><div>Collaborative robots (cobots) enhanced by artificial intelligence (AI) are enabling intelligent, human-centric manufacturing environments. These dynamic settings require cobots with cognitive intelligence, i.e., capabilities covering perception, learning, decision-making, and adaptation. Such intelligence enables proactive collaboration that integrates bidirectional instructional and cooperative competencies. However, while extensive research has focused on improving the performance of robot collaborative skills, systematic investigations into the instructional capabilities of cobots remain notably limited. To lay the technological foundation for addressing this gap, this survey adopts a multimodal perspective to review three essential aspects of this field: (1) robot learning from video (LfV) for instructional capabilities acquisition, (2) robot-guided instruction methodologies, and (3) feedback-driven adaptation. We present a systematic review of technologies for representing human actions, object states, and human-object interactions (HOI), with a particular focus on multimodal data sources from video. Furthermore, we analyze diverse instructional strategies, including visual guidance, auditory directives, robot-performed actions, emphasizing their effectiveness in robot-guided instruction. A significant focus is placed on feedback-driven adaptation mechanisms, which enable cobots to dynamically refine their instructional capabilities based on user feedback. We identify key challenges such as environmental complexity, user variability, real-time processing constraints, and trust-building requirements, while also highlighting emerging opportunities in multimodal integration, AI-powered robots, and collaborative learning systems. Finally, we underscore the transformative potential of instructional cobots in smart manufacturing and emphasize the necessity for further research.</div></div>","PeriodicalId":16227,"journal":{"name":"Journal of Manufacturing Systems","volume":"83 ","pages":"Pages 29-45"},"PeriodicalIF":14.2000,"publicationDate":"2025-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Manufacturing Systems","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0278612525002171","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, INDUSTRIAL","Score":null,"Total":0}
引用次数: 0
Abstract
Collaborative robots (cobots) enhanced by artificial intelligence (AI) are enabling intelligent, human-centric manufacturing environments. These dynamic settings require cobots with cognitive intelligence, i.e., capabilities covering perception, learning, decision-making, and adaptation. Such intelligence enables proactive collaboration that integrates bidirectional instructional and cooperative competencies. However, while extensive research has focused on improving the performance of robot collaborative skills, systematic investigations into the instructional capabilities of cobots remain notably limited. To lay the technological foundation for addressing this gap, this survey adopts a multimodal perspective to review three essential aspects of this field: (1) robot learning from video (LfV) for instructional capabilities acquisition, (2) robot-guided instruction methodologies, and (3) feedback-driven adaptation. We present a systematic review of technologies for representing human actions, object states, and human-object interactions (HOI), with a particular focus on multimodal data sources from video. Furthermore, we analyze diverse instructional strategies, including visual guidance, auditory directives, robot-performed actions, emphasizing their effectiveness in robot-guided instruction. A significant focus is placed on feedback-driven adaptation mechanisms, which enable cobots to dynamically refine their instructional capabilities based on user feedback. We identify key challenges such as environmental complexity, user variability, real-time processing constraints, and trust-building requirements, while also highlighting emerging opportunities in multimodal integration, AI-powered robots, and collaborative learning systems. Finally, we underscore the transformative potential of instructional cobots in smart manufacturing and emphasize the necessity for further research.
期刊介绍:
The Journal of Manufacturing Systems is dedicated to showcasing cutting-edge fundamental and applied research in manufacturing at the systems level. Encompassing products, equipment, people, information, control, and support functions, manufacturing systems play a pivotal role in the economical and competitive development, production, delivery, and total lifecycle of products, meeting market and societal needs.
With a commitment to publishing archival scholarly literature, the journal strives to advance the state of the art in manufacturing systems and foster innovation in crafting efficient, robust, and sustainable manufacturing systems. The focus extends from equipment-level considerations to the broader scope of the extended enterprise. The Journal welcomes research addressing challenges across various scales, including nano, micro, and macro-scale manufacturing, and spanning diverse sectors such as aerospace, automotive, energy, and medical device manufacturing.