海报摘要:动态实时微传感器运动捕捉

Shuai Huang, Shuyan Sun, Zhipei Huang, Jiankang Wu, X. Meng, Guanhong Tao, N. Zhang, Li Yang
{"title":"海报摘要:动态实时微传感器运动捕捉","authors":"Shuai Huang, Shuyan Sun, Zhipei Huang, Jiankang Wu, X. Meng, Guanhong Tao, N. Zhang, Li Yang","doi":"10.1109/IPSN.2012.6920977","DOIUrl":null,"url":null,"abstract":"Commercial optical human motion capture systems perform well in studio-like environments, but they do not provide solution in daily-life surroundings. Micro-sensor motion capture has shown its potentials because of its ubiquity and low cost. We present an ambulatory low-cost real-time motion capture system using wearable micro-sensors (accelerometers, magnetometers and gyroscopes), which can capture and reconstruct human motion in real-time almost every-where. It mainly consists of three parts: a sensor subsystem, a data fusion subsystem and an animation subsystem. The sensor subsystem collects human motion signals and transfers them into the data fusion subsystem. The data fusion subsystem performs sensor fusion to obtain motion information, i.e., the orientation and position of each body segment. Using the motion information from the data fusion subsystem, the animation subsystem drives the avatar in the 3D virtual world in order to reconstruct human motion. All the processes are accomplished in real-time. The experimental results show that our system can capture motions and drive animations in real-time vividly without drift and delay. And the output from our system can be made use of in film-making, sports training and argument reality applications, etc.","PeriodicalId":231003,"journal":{"name":"2012 ACM/IEEE 11th International Conference on Information Processing in Sensor Networks (IPSN)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2012-04-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Poster abstract: Ambulatory real-time micro-sensor motion capture\",\"authors\":\"Shuai Huang, Shuyan Sun, Zhipei Huang, Jiankang Wu, X. Meng, Guanhong Tao, N. Zhang, Li Yang\",\"doi\":\"10.1109/IPSN.2012.6920977\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Commercial optical human motion capture systems perform well in studio-like environments, but they do not provide solution in daily-life surroundings. Micro-sensor motion capture has shown its potentials because of its ubiquity and low cost. We present an ambulatory low-cost real-time motion capture system using wearable micro-sensors (accelerometers, magnetometers and gyroscopes), which can capture and reconstruct human motion in real-time almost every-where. It mainly consists of three parts: a sensor subsystem, a data fusion subsystem and an animation subsystem. The sensor subsystem collects human motion signals and transfers them into the data fusion subsystem. The data fusion subsystem performs sensor fusion to obtain motion information, i.e., the orientation and position of each body segment. Using the motion information from the data fusion subsystem, the animation subsystem drives the avatar in the 3D virtual world in order to reconstruct human motion. All the processes are accomplished in real-time. The experimental results show that our system can capture motions and drive animations in real-time vividly without drift and delay. And the output from our system can be made use of in film-making, sports training and argument reality applications, etc.\",\"PeriodicalId\":231003,\"journal\":{\"name\":\"2012 ACM/IEEE 11th International Conference on Information Processing in Sensor Networks (IPSN)\",\"volume\":\"37 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2012-04-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2012 ACM/IEEE 11th International Conference on Information Processing in Sensor Networks (IPSN)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IPSN.2012.6920977\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2012 ACM/IEEE 11th International Conference on Information Processing in Sensor Networks (IPSN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IPSN.2012.6920977","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

摘要

商业光学人体动作捕捉系统在类似工作室的环境中表现良好,但它们不能提供在日常生活环境中的解决方案。微传感器运动捕捉因其普遍存在和低成本而显示出其潜力。我们提出了一种使用可穿戴微传感器(加速度计、磁力计和陀螺仪)的低成本动态实时运动捕捉系统,该系统几乎可以在任何地方实时捕获和重建人体运动。该系统主要由三部分组成:传感器子系统、数据融合子系统和动画子系统。传感器子系统采集人体运动信号,并将其传输到数据融合子系统。数据融合子系统通过传感器融合获得运动信息,即每个身体部分的方向和位置。动画子系统利用数据融合子系统的运动信息,驱动虚拟人物在三维虚拟世界中进行人体运动重构。所有的过程都是实时完成的。实验结果表明,该系统能够实时逼真地捕捉运动和驱动动画,无漂移和延迟。该系统的输出可用于电影制作、体育训练和辩论现实应用等。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Poster abstract: Ambulatory real-time micro-sensor motion capture
Commercial optical human motion capture systems perform well in studio-like environments, but they do not provide solution in daily-life surroundings. Micro-sensor motion capture has shown its potentials because of its ubiquity and low cost. We present an ambulatory low-cost real-time motion capture system using wearable micro-sensors (accelerometers, magnetometers and gyroscopes), which can capture and reconstruct human motion in real-time almost every-where. It mainly consists of three parts: a sensor subsystem, a data fusion subsystem and an animation subsystem. The sensor subsystem collects human motion signals and transfers them into the data fusion subsystem. The data fusion subsystem performs sensor fusion to obtain motion information, i.e., the orientation and position of each body segment. Using the motion information from the data fusion subsystem, the animation subsystem drives the avatar in the 3D virtual world in order to reconstruct human motion. All the processes are accomplished in real-time. The experimental results show that our system can capture motions and drive animations in real-time vividly without drift and delay. And the output from our system can be made use of in film-making, sports training and argument reality applications, etc.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信