智能手势识别提高设计效率智能多式联运系统

S. Chhabria, R. Dharaskar, V. Thakare
{"title":"智能手势识别提高设计效率智能多式联运系统","authors":"S. Chhabria, R. Dharaskar, V. Thakare","doi":"10.1109/ICETET.2013.70","DOIUrl":null,"url":null,"abstract":"Technologies used in multimodal interfaces include conventional direct-manipulation devices like the keyboard, mouse, and pen and touch screen, as well as progressively more advanced recognition technologies such as speech recognition, 2D and 3D gesture recognition, and lip movement and gaze tracking. Usability studies, exploring and evaluating the human factors involved in multimodal input, provide useful insight and guidance toward the design and implementation of multimodal interfaces. The primary goal in the design of any user interface is to facilitate the interaction between user and machine. This user-centered goal is the guiding force behind choices made in the design process. There are, of course, many system engineering issues that influence interface design decisions such as schedules, proper functionality, reliability, etc. However, addressing these issues ideally serves the purpose of creating a better user experience with the system. One purpose of researching multimodal interfaces from an HCI perspective is to evaluate how to take advantage of the benefits they provide over unimodal recognition-based interfaces and conventional keyboard and mouse interfaces. Such advantages include flexibility, availability, adaptability, efficiency, lower error rate, and a more intuitive and natural interaction [1, 2]. Technology is one way to remove the hindrances and benefit the people. It is found that all the recent techniques for Multimodal System are based on computer tools, techniques and technologies like: Artificial Intelligence; Expert System; Graph Matching; Soft Computing; Natural Language Processing (NLP); Multi-Modal and HCI. Several researchers have explored their possibilities and have achieved results to certain extent, but progress in Speech Recognition & Language Understanding using gestures, & optimization as a whole has been limited which gives us future scope to go ahead in this domain.","PeriodicalId":440967,"journal":{"name":"2013 6th International Conference on Emerging Trends in Engineering and Technology","volume":"23 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2013-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Intelligent Gesture Recognition to Design more Efficient & Intelligent Multimodal System\",\"authors\":\"S. Chhabria, R. Dharaskar, V. Thakare\",\"doi\":\"10.1109/ICETET.2013.70\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Technologies used in multimodal interfaces include conventional direct-manipulation devices like the keyboard, mouse, and pen and touch screen, as well as progressively more advanced recognition technologies such as speech recognition, 2D and 3D gesture recognition, and lip movement and gaze tracking. Usability studies, exploring and evaluating the human factors involved in multimodal input, provide useful insight and guidance toward the design and implementation of multimodal interfaces. The primary goal in the design of any user interface is to facilitate the interaction between user and machine. This user-centered goal is the guiding force behind choices made in the design process. There are, of course, many system engineering issues that influence interface design decisions such as schedules, proper functionality, reliability, etc. However, addressing these issues ideally serves the purpose of creating a better user experience with the system. One purpose of researching multimodal interfaces from an HCI perspective is to evaluate how to take advantage of the benefits they provide over unimodal recognition-based interfaces and conventional keyboard and mouse interfaces. Such advantages include flexibility, availability, adaptability, efficiency, lower error rate, and a more intuitive and natural interaction [1, 2]. Technology is one way to remove the hindrances and benefit the people. It is found that all the recent techniques for Multimodal System are based on computer tools, techniques and technologies like: Artificial Intelligence; Expert System; Graph Matching; Soft Computing; Natural Language Processing (NLP); Multi-Modal and HCI. Several researchers have explored their possibilities and have achieved results to certain extent, but progress in Speech Recognition & Language Understanding using gestures, & optimization as a whole has been limited which gives us future scope to go ahead in this domain.\",\"PeriodicalId\":440967,\"journal\":{\"name\":\"2013 6th International Conference on Emerging Trends in Engineering and Technology\",\"volume\":\"23 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2013-12-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2013 6th International Conference on Emerging Trends in Engineering and Technology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICETET.2013.70\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2013 6th International Conference on Emerging Trends in Engineering and Technology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICETET.2013.70","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

多模态界面中使用的技术包括传统的直接操作设备,如键盘、鼠标、笔和触摸屏,以及越来越先进的识别技术,如语音识别、2D和3D手势识别、嘴唇运动和凝视跟踪。可用性研究,探索和评估涉及多模态输入的人为因素,为多模态界面的设计和实现提供有用的见解和指导。任何用户界面设计的主要目标都是促进用户和机器之间的交互。这种以用户为中心的目标是设计过程中做出选择的指导力量。当然,有许多系统工程问题会影响接口设计决策,如进度、适当的功能、可靠性等。然而,理想地解决这些问题有助于为系统创建更好的用户体验。从人机交互的角度研究多模态接口的一个目的是评估如何利用它们提供的优于基于单模态识别的接口和传统的键盘和鼠标接口的优势。这些优势包括灵活性、可用性、适应性、效率、更低的错误率以及更直观和自然的交互[1,2]。技术是消除障碍、造福人民的一种方式。研究发现,最近的多式联运系统技术都是基于计算机工具、技术和技术,如:人工智能;专家系统;图像匹配;软计算;自然语言处理;多模式和HCI。一些研究人员已经探索了它们的可能性,并在一定程度上取得了成果,但在使用手势的语音识别和语言理解以及整体优化方面的进展有限,这给了我们未来在这一领域继续前进的空间。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Intelligent Gesture Recognition to Design more Efficient & Intelligent Multimodal System
Technologies used in multimodal interfaces include conventional direct-manipulation devices like the keyboard, mouse, and pen and touch screen, as well as progressively more advanced recognition technologies such as speech recognition, 2D and 3D gesture recognition, and lip movement and gaze tracking. Usability studies, exploring and evaluating the human factors involved in multimodal input, provide useful insight and guidance toward the design and implementation of multimodal interfaces. The primary goal in the design of any user interface is to facilitate the interaction between user and machine. This user-centered goal is the guiding force behind choices made in the design process. There are, of course, many system engineering issues that influence interface design decisions such as schedules, proper functionality, reliability, etc. However, addressing these issues ideally serves the purpose of creating a better user experience with the system. One purpose of researching multimodal interfaces from an HCI perspective is to evaluate how to take advantage of the benefits they provide over unimodal recognition-based interfaces and conventional keyboard and mouse interfaces. Such advantages include flexibility, availability, adaptability, efficiency, lower error rate, and a more intuitive and natural interaction [1, 2]. Technology is one way to remove the hindrances and benefit the people. It is found that all the recent techniques for Multimodal System are based on computer tools, techniques and technologies like: Artificial Intelligence; Expert System; Graph Matching; Soft Computing; Natural Language Processing (NLP); Multi-Modal and HCI. Several researchers have explored their possibilities and have achieved results to certain extent, but progress in Speech Recognition & Language Understanding using gestures, & optimization as a whole has been limited which gives us future scope to go ahead in this domain.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信