面向机器人控制的鲁棒自我中心手势分析

Hongyong Song, Weijiang Feng, Naiyang Guan, Xuhui Huang, Zhigang Luo
{"title":"面向机器人控制的鲁棒自我中心手势分析","authors":"Hongyong Song, Weijiang Feng, Naiyang Guan, Xuhui Huang, Zhigang Luo","doi":"10.1109/SIPROCESS.2016.7888345","DOIUrl":null,"url":null,"abstract":"Wearable device with an ego-centric camera would be the next generation device for human-computer interaction such as robot control. Hand gesture is a natural way of ego-centric human-computer interaction. In this paper, we present an ego-centric multi-stage hand gesture analysis pipeline for robot control which works robustly in the unconstrained environment with varying luminance. In particular, we first propose an adaptive color and contour based hand segmentation method to segment hand region from the ego-centric viewpoint. We then propose a convex U-shaped curve detection algorithm to precisely detect positions of fingertips. And parallelly, we utilize the convolutional neural networks to recognize hand gestures. Based on these techniques, we combine most information of hand to control the robot and develop a hand gesture analysis system on an iPhone and a robot arm platform to validate its effectiveness. Experimental result demonstrates that our method works perfectly on controlling the robot arm by hand gesture in real time.","PeriodicalId":142802,"journal":{"name":"2016 IEEE International Conference on Signal and Image Processing (ICSIP)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":"{\"title\":\"Towards robust ego-centric hand gesture analysis for robot control\",\"authors\":\"Hongyong Song, Weijiang Feng, Naiyang Guan, Xuhui Huang, Zhigang Luo\",\"doi\":\"10.1109/SIPROCESS.2016.7888345\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Wearable device with an ego-centric camera would be the next generation device for human-computer interaction such as robot control. Hand gesture is a natural way of ego-centric human-computer interaction. In this paper, we present an ego-centric multi-stage hand gesture analysis pipeline for robot control which works robustly in the unconstrained environment with varying luminance. In particular, we first propose an adaptive color and contour based hand segmentation method to segment hand region from the ego-centric viewpoint. We then propose a convex U-shaped curve detection algorithm to precisely detect positions of fingertips. And parallelly, we utilize the convolutional neural networks to recognize hand gestures. Based on these techniques, we combine most information of hand to control the robot and develop a hand gesture analysis system on an iPhone and a robot arm platform to validate its effectiveness. Experimental result demonstrates that our method works perfectly on controlling the robot arm by hand gesture in real time.\",\"PeriodicalId\":142802,\"journal\":{\"name\":\"2016 IEEE International Conference on Signal and Image Processing (ICSIP)\",\"volume\":\"6 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2016-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"5\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2016 IEEE International Conference on Signal and Image Processing (ICSIP)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SIPROCESS.2016.7888345\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 IEEE International Conference on Signal and Image Processing (ICSIP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SIPROCESS.2016.7888345","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5

摘要

带有以自我为中心的摄像头的可穿戴设备将成为机器人控制等人机交互的下一代设备。手势是一种以自我为中心的自然人机交互方式。本文提出了一种以自我为中心的多阶段手势分析流水线,该流水线能够在无约束的变亮度环境中鲁棒地工作。特别地,我们首先提出了一种基于自适应颜色和轮廓的手部分割方法,从自我中心的角度对手部区域进行分割。然后,我们提出了一种凸u形曲线检测算法来精确检测指尖的位置。同时,我们利用卷积神经网络来识别手势。在此基础上,我们结合手的大部分信息来控制机器人,并在iPhone和机械臂平台上开发了手势分析系统来验证其有效性。实验结果表明,该方法可以很好地实现对机械臂的实时手势控制。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Towards robust ego-centric hand gesture analysis for robot control
Wearable device with an ego-centric camera would be the next generation device for human-computer interaction such as robot control. Hand gesture is a natural way of ego-centric human-computer interaction. In this paper, we present an ego-centric multi-stage hand gesture analysis pipeline for robot control which works robustly in the unconstrained environment with varying luminance. In particular, we first propose an adaptive color and contour based hand segmentation method to segment hand region from the ego-centric viewpoint. We then propose a convex U-shaped curve detection algorithm to precisely detect positions of fingertips. And parallelly, we utilize the convolutional neural networks to recognize hand gestures. Based on these techniques, we combine most information of hand to control the robot and develop a hand gesture analysis system on an iPhone and a robot arm platform to validate its effectiveness. Experimental result demonstrates that our method works perfectly on controlling the robot arm by hand gesture in real time.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信