Liam Roy;Elizabeth A. Croft;Alex Ramirez;Dana Kulić
{"title":"GPT-Driven Gestures: Leveraging Large Language Models to Generate Expressive Robot Motion for Enhanced Human-Robot Interaction","authors":"Liam Roy;Elizabeth A. Croft;Alex Ramirez;Dana Kulić","doi":"10.1109/LRA.2025.3547631","DOIUrl":null,"url":null,"abstract":"Expressive robot motion is a form of nonverbal communication that enables robots to convey their internal states, fostering effective human-robot interaction. A key step in designing expressive robot motions is developing a mapping from the desired states the robot will express to the robot's hardware and available degrees of freedom (design space). This letter introduces a novel framework to autonomously generate this mapping by leveraging a large language model (LLM) to select motion parameters and their values for target robot states. We evaluate expressive robot body language displayed on a Unitree Go1 quadruped as generated by a Generative Pre-trained Transformer (GPT) provided with a set of adjustable motion parameters. Through a two-part study (N = 120), we compared LLM-generated expressive motions with both randomly selected and human-selected expressions. Our results show that participants viewing LLM-generated expressions achieve a significantly higher state classification accuracy over random baselines and perform comparably with human-generated expressions. Additionally, in our post-hoc analysis we find that the Earth Movers Distance provides a useful metric for identifying similar expressions in the design space that lead to classification confusion.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 5","pages":"4172-4179"},"PeriodicalIF":4.6000,"publicationDate":"2025-03-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Robotics and Automation Letters","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10909198/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ROBOTICS","Score":null,"Total":0}
引用次数: 0
Abstract
Expressive robot motion is a form of nonverbal communication that enables robots to convey their internal states, fostering effective human-robot interaction. A key step in designing expressive robot motions is developing a mapping from the desired states the robot will express to the robot's hardware and available degrees of freedom (design space). This letter introduces a novel framework to autonomously generate this mapping by leveraging a large language model (LLM) to select motion parameters and their values for target robot states. We evaluate expressive robot body language displayed on a Unitree Go1 quadruped as generated by a Generative Pre-trained Transformer (GPT) provided with a set of adjustable motion parameters. Through a two-part study (N = 120), we compared LLM-generated expressive motions with both randomly selected and human-selected expressions. Our results show that participants viewing LLM-generated expressions achieve a significantly higher state classification accuracy over random baselines and perform comparably with human-generated expressions. Additionally, in our post-hoc analysis we find that the Earth Movers Distance provides a useful metric for identifying similar expressions in the design space that lead to classification confusion.
期刊介绍:
The scope of this journal is to publish peer-reviewed articles that provide a timely and concise account of innovative research ideas and application results, reporting significant theoretical findings and application case studies in areas of robotics and automation.