基于深度卷积神经网络的手语手指拼写识别系统

Antara Howal, Atharva Golapkar, Yunus Khan, Siddhantha Bokade, S. Varma, Madhura Vyawahare
{"title":"基于深度卷积神经网络的手语手指拼写识别系统","authors":"Antara Howal, Atharva Golapkar, Yunus Khan, Siddhantha Bokade, S. Varma, Madhura Vyawahare","doi":"10.1109/ICNTE56631.2023.10146675","DOIUrl":null,"url":null,"abstract":"Sign Language is used by deaf and hard hearing people as it is the efficient way to communicate among themselves as well as with other people. The demand for an automated system, which converts sign language into a normal message and vice versa, has been increased to provide a normal life to people with hearing difficulties. Each sign language has a distinct character with variations in the shape of the hand, the profile of movement, and the position of the hand, face, and body parts that contribute to each character. Pattern recognition and gesture recognition are majorly used technologies for resolving the problem. In this work, real-time American Sign Language is carried out for finger-spelling recognition using Deep Convolutional Neural Network for the classification of data to bridge the gap of communication between hearing disabled people and normal people. A new dataset of more than 50 thousand images is prepared for training purposes. Outputs are based on inputs given by the user, while the designed interface provides the end user with alphabets A to Z and Numerals from 0 to 9 prediction. The overall performance based on the proposed approach is being evaluated by the use of a dataset of real-depth images captured from users. Accuracy of the system is 99.70%.","PeriodicalId":158124,"journal":{"name":"2023 5th Biennial International Conference on Nascent Technologies in Engineering (ICNTE)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Sign Language Finger-Spelling Recognition System Using Deep Convolutional Neural Network\",\"authors\":\"Antara Howal, Atharva Golapkar, Yunus Khan, Siddhantha Bokade, S. Varma, Madhura Vyawahare\",\"doi\":\"10.1109/ICNTE56631.2023.10146675\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Sign Language is used by deaf and hard hearing people as it is the efficient way to communicate among themselves as well as with other people. The demand for an automated system, which converts sign language into a normal message and vice versa, has been increased to provide a normal life to people with hearing difficulties. Each sign language has a distinct character with variations in the shape of the hand, the profile of movement, and the position of the hand, face, and body parts that contribute to each character. Pattern recognition and gesture recognition are majorly used technologies for resolving the problem. In this work, real-time American Sign Language is carried out for finger-spelling recognition using Deep Convolutional Neural Network for the classification of data to bridge the gap of communication between hearing disabled people and normal people. A new dataset of more than 50 thousand images is prepared for training purposes. Outputs are based on inputs given by the user, while the designed interface provides the end user with alphabets A to Z and Numerals from 0 to 9 prediction. The overall performance based on the proposed approach is being evaluated by the use of a dataset of real-depth images captured from users. Accuracy of the system is 99.70%.\",\"PeriodicalId\":158124,\"journal\":{\"name\":\"2023 5th Biennial International Conference on Nascent Technologies in Engineering (ICNTE)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-01-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 5th Biennial International Conference on Nascent Technologies in Engineering (ICNTE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICNTE56631.2023.10146675\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 5th Biennial International Conference on Nascent Technologies in Engineering (ICNTE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICNTE56631.2023.10146675","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

聋哑人和重听人使用手语,因为这是他们之间以及与他人交流的有效方式。为了让听力有障碍的人过上正常的生活,对将手语转换为正常信息的自动化系统的需求有所增加。每一种手语都有一个独特的特征,手的形状、运动的轮廓、手的位置、脸和身体部位都有不同的特征。模式识别和手势识别是解决这一问题的主要技术。本研究利用深度卷积神经网络(Deep Convolutional Neural Network)对数据进行分类,进行实时美式手语的拼写识别,以弥补听障人士与正常人之间交流的差距。为训练目的准备了一个超过5万张图像的新数据集。输出基于用户给出的输入,而设计的界面为最终用户提供字母A到Z和数字0到9的预测。通过使用从用户处捕获的真实深度图像数据集来评估基于所提出方法的总体性能。系统的准确率为99.70%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Sign Language Finger-Spelling Recognition System Using Deep Convolutional Neural Network
Sign Language is used by deaf and hard hearing people as it is the efficient way to communicate among themselves as well as with other people. The demand for an automated system, which converts sign language into a normal message and vice versa, has been increased to provide a normal life to people with hearing difficulties. Each sign language has a distinct character with variations in the shape of the hand, the profile of movement, and the position of the hand, face, and body parts that contribute to each character. Pattern recognition and gesture recognition are majorly used technologies for resolving the problem. In this work, real-time American Sign Language is carried out for finger-spelling recognition using Deep Convolutional Neural Network for the classification of data to bridge the gap of communication between hearing disabled people and normal people. A new dataset of more than 50 thousand images is prepared for training purposes. Outputs are based on inputs given by the user, while the designed interface provides the end user with alphabets A to Z and Numerals from 0 to 9 prediction. The overall performance based on the proposed approach is being evaluated by the use of a dataset of real-depth images captured from users. Accuracy of the system is 99.70%.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信