{"title":"Investigation of Joint Action in Go/No-Go Tasks: Development of a Human-Like Eye Robot and Verification of Action Space","authors":"Kotaro Hayashi","doi":"10.1007/s12369-024-01168-4","DOIUrl":null,"url":null,"abstract":"<p>Human–robot collaboration (HRC) is a natural progression of technological development and can improve job performance, address labor shortages, and reduce labor costs. However, it is still uncertain whether joint action, similar to that occurring between humans, can be replicated between humans and robots. Many robotic researchers have focused on joint action, and it has been demonstrated that gaze cueing plays a significant role in this context. Currently, previous studies on joint action use humanoids; however, robots utilized in the research on human-robot collaboration lack human-like eyes needed for verification. Therefore, this study focuses on the development of an eye robot with gaze-cueing behaviors that can be easily integrated into existing robotic systems. As another theme of this study, we proposed the use of fixation duration as a new metric, which is distinct from the commonly used response time, for the quantitative evaluation of joint action research. These are verified through a Go/No-go task under six conditions—three behavioral (i.e., joint action, joint attention-only, and alone), each with two partner conditions (robot or human partner). While developing a human-like eye robot, this study demonstrates the potential of a robot to be a better joint action partner than an uncertain human, with participants exhibiting the best reaction times when partnered with a robot. The shared action space of the participants was investigated, where a transference of the action space indicates the expression of joint action. The fixation duration indicates that the proposed robot cause participants to move their action space to include that of the robot. These results suggest that the proposed collaborative robot can initiate a joint action between a robot and a human, and can perform as a more effective partner in joint actions compared to an unfamiliar human. This study showcased the capacity of fixation duration as a quantitative assessment metric for joint action.</p>","PeriodicalId":14361,"journal":{"name":"International Journal of Social Robotics","volume":"26 1","pages":""},"PeriodicalIF":3.8000,"publicationDate":"2024-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Social Robotics","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s12369-024-01168-4","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ROBOTICS","Score":null,"Total":0}
引用次数: 0
Abstract
Human–robot collaboration (HRC) is a natural progression of technological development and can improve job performance, address labor shortages, and reduce labor costs. However, it is still uncertain whether joint action, similar to that occurring between humans, can be replicated between humans and robots. Many robotic researchers have focused on joint action, and it has been demonstrated that gaze cueing plays a significant role in this context. Currently, previous studies on joint action use humanoids; however, robots utilized in the research on human-robot collaboration lack human-like eyes needed for verification. Therefore, this study focuses on the development of an eye robot with gaze-cueing behaviors that can be easily integrated into existing robotic systems. As another theme of this study, we proposed the use of fixation duration as a new metric, which is distinct from the commonly used response time, for the quantitative evaluation of joint action research. These are verified through a Go/No-go task under six conditions—three behavioral (i.e., joint action, joint attention-only, and alone), each with two partner conditions (robot or human partner). While developing a human-like eye robot, this study demonstrates the potential of a robot to be a better joint action partner than an uncertain human, with participants exhibiting the best reaction times when partnered with a robot. The shared action space of the participants was investigated, where a transference of the action space indicates the expression of joint action. The fixation duration indicates that the proposed robot cause participants to move their action space to include that of the robot. These results suggest that the proposed collaborative robot can initiate a joint action between a robot and a human, and can perform as a more effective partner in joint actions compared to an unfamiliar human. This study showcased the capacity of fixation duration as a quantitative assessment metric for joint action.
期刊介绍:
Social Robotics is the study of robots that are able to interact and communicate among themselves, with humans, and with the environment, within the social and cultural structure attached to its role. The journal covers a broad spectrum of topics related to the latest technologies, new research results and developments in the area of social robotics on all levels, from developments in core enabling technologies to system integration, aesthetic design, applications and social implications. It provides a platform for like-minded researchers to present their findings and latest developments in social robotics, covering relevant advances in engineering, computing, arts and social sciences.
The journal publishes original, peer reviewed articles and contributions on innovative ideas and concepts, new discoveries and improvements, as well as novel applications, by leading researchers and developers regarding the latest fundamental advances in the core technologies that form the backbone of social robotics, distinguished developmental projects in the area, as well as seminal works in aesthetic design, ethics and philosophy, studies on social impact and influence, pertaining to social robotics.