Integrating joint and surface for human action recognition in indoor environments

Qingyang Li, Yu Zhou, Anlong Ming
{"title":"Integrating joint and surface for human action recognition in indoor environments","authors":"Qingyang Li, Yu Zhou, Anlong Ming","doi":"10.1109/SPAC.2014.6982665","DOIUrl":null,"url":null,"abstract":"Action recognition has a long research history, despite several contributed approaches have been introduced, it remains a challenging task in computer vision. In this paper, we present a uniform fusion framework for action recognition, which integrates not only the local depth cues but also the global depth cues. Firstly, the action recognition task is formulated as the maximize the posterior probability, and then the observation for the original action is decomposed into the sub-observations for each individual feature representation strategy of the original action. For the local depth cues, the joints inside the human skeleton is employed to model the local variation of the human motion. In addition, the normal of the depth surface is utilized as the global cue to capture the holistic structure of the human motion. Rather than using the original feature directly, the support vector machine model learning both the discriminative local cue (i.e., the joint) and the discriminative global cue (i.e., the depth surface), respectively. The presented approach is validated on the famous MSR Daily Activity 3D Dataset. And the experimental results demonstrate that our fusion approach can outperform the baseline approaches.","PeriodicalId":326246,"journal":{"name":"Proceedings 2014 IEEE International Conference on Security, Pattern Analysis, and Cybernetics (SPAC)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2014-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings 2014 IEEE International Conference on Security, Pattern Analysis, and Cybernetics (SPAC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SPAC.2014.6982665","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

Action recognition has a long research history, despite several contributed approaches have been introduced, it remains a challenging task in computer vision. In this paper, we present a uniform fusion framework for action recognition, which integrates not only the local depth cues but also the global depth cues. Firstly, the action recognition task is formulated as the maximize the posterior probability, and then the observation for the original action is decomposed into the sub-observations for each individual feature representation strategy of the original action. For the local depth cues, the joints inside the human skeleton is employed to model the local variation of the human motion. In addition, the normal of the depth surface is utilized as the global cue to capture the holistic structure of the human motion. Rather than using the original feature directly, the support vector machine model learning both the discriminative local cue (i.e., the joint) and the discriminative global cue (i.e., the depth surface), respectively. The presented approach is validated on the famous MSR Daily Activity 3D Dataset. And the experimental results demonstrate that our fusion approach can outperform the baseline approaches.
结合关节和表面进行室内环境人体动作识别
动作识别有着悠久的研究历史,尽管已经引入了一些有贡献的方法,但它仍然是计算机视觉中的一项具有挑战性的任务。本文提出了一种统一的动作识别融合框架,既融合了局部深度线索,又融合了全局深度线索。首先将动作识别任务表述为后验概率最大化,然后将原始动作的观测值分解为原始动作各个特征表示策略的子观测值。对于局部深度线索,利用人体骨骼内部的关节来模拟人体运动的局部变化。此外,利用深度表面的法线作为全局线索来捕捉人体运动的整体结构。支持向量机模型不是直接使用原始特征,而是分别学习判别性局部线索(即关节)和判别性全局线索(即深度面)。该方法在著名的MSR日常活动3D数据集上进行了验证。实验结果表明,我们的融合方法优于基线方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信