Action similarity judgment based on kinematic primitives

Vipul Nair, Paul E. Hemeren, Alessia Vignolo, Nicoletta Noceti, Elena Nicora, A. Sciutti, F. Rea, E. Billing, F. Odone, G. Sandini
{"title":"Action similarity judgment based on kinematic primitives","authors":"Vipul Nair, Paul E. Hemeren, Alessia Vignolo, Nicoletta Noceti, Elena Nicora, A. Sciutti, F. Rea, E. Billing, F. Odone, G. Sandini","doi":"10.1109/ICDL-EpiRob48136.2020.9278047","DOIUrl":null,"url":null,"abstract":"Understanding which features humans rely on - in visually recognizing action similarity is a crucial step towards a clearer picture of human action perception from a learning and developmental perspective. In the present work, we investigate to which extent a computational model based on kinematics can determine action similarity and how its performance relates to human similarity judgments of the same actions. To this aim, twelve participants perform an action similarity task, and their performances are compared to that of a computational model solving the same task. The chosen model has its roots in developmental robotics and performs action classification based on learned kinematic primitives. The comparative experiment results show that both the model and human participants can reliably identify whether two actions are the same or not. However, the model produces more false hits and has a greater selection bias than human participants. A possible reason for this is the particular sensitivity of the model towards kinematic primitives of the presented actions. In a second experiment, human participants' performance on an action identification task indicated that they relied solely on kinematic information rather than on action semantics. The results show that both the model and human performance are highly accurate in an action similarity task based on kinematic-level features, which can provide an essential basis for classifying human actions.","PeriodicalId":114948,"journal":{"name":"2020 Joint IEEE 10th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob)","volume":"76 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 Joint IEEE 10th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDL-EpiRob48136.2020.9278047","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

Abstract

Understanding which features humans rely on - in visually recognizing action similarity is a crucial step towards a clearer picture of human action perception from a learning and developmental perspective. In the present work, we investigate to which extent a computational model based on kinematics can determine action similarity and how its performance relates to human similarity judgments of the same actions. To this aim, twelve participants perform an action similarity task, and their performances are compared to that of a computational model solving the same task. The chosen model has its roots in developmental robotics and performs action classification based on learned kinematic primitives. The comparative experiment results show that both the model and human participants can reliably identify whether two actions are the same or not. However, the model produces more false hits and has a greater selection bias than human participants. A possible reason for this is the particular sensitivity of the model towards kinematic primitives of the presented actions. In a second experiment, human participants' performance on an action identification task indicated that they relied solely on kinematic information rather than on action semantics. The results show that both the model and human performance are highly accurate in an action similarity task based on kinematic-level features, which can provide an essential basis for classifying human actions.
基于运动学原语的动作相似度判断
从学习和发展的角度来看,理解人类依赖于哪些特征——在视觉上识别动作相似性,是朝着更清晰地了解人类动作感知迈出的关键一步。在目前的工作中,我们研究了基于运动学的计算模型在多大程度上可以确定动作相似性,以及它的性能如何与人类对相同动作的相似性判断相关联。为此,12名参与者执行一个动作相似性任务,并将他们的表现与解决相同任务的计算模型的表现进行比较。所选择的模型植根于发展机器人,并基于学习到的运动学原语进行动作分类。对比实验结果表明,该模型和人类参与者都能可靠地识别两个动作是否相同。然而,与人类参与者相比,该模型产生了更多的错误命中,并且具有更大的选择偏差。一个可能的原因是该模型对所呈现动作的运动学原语的特殊敏感性。在第二个实验中,人类参与者在动作识别任务上的表现表明,他们完全依赖于运动学信息,而不是动作语义。结果表明,在基于运动级特征的动作相似任务中,该模型和人的表现都具有较高的准确率,为人类动作分类提供了必要的依据。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信