Real-Time Sign Language Converter for Mute and Deaf People

Akshit J. Dhruv, Santosh Kumar Bharti
{"title":"Real-Time Sign Language Converter for Mute and Deaf People","authors":"Akshit J. Dhruv, Santosh Kumar Bharti","doi":"10.1109/aimv53313.2021.9670928","DOIUrl":null,"url":null,"abstract":"Deaf people may get irritated due to the problem of not being able to share their views with common people, which may affect their day-to-day life. This is the main reason to develop such system that can help these people and they can also put their thoughts forward similar to other people who don’t have such problem. The advancement in the Artificial intelligence provides the door for developing the system that overcome this difficulty. So this project aims on developing a system which will be able to convert the speech to text for the deaf person, and also sometimes the person might not be able to understand just by text, so the speech will also get converted to the universal sign language. Similarly, for the mute people the sign language which they are using will get converted to speech. We will take help of various ML and AI concepts along with NLP to develop the accurate model. Convolutional neural networks (CNN) will be used for prediction as it is efficient in predicting image input, also as lip movements are fast and continuous so it is hard to capture so along with CNN, the use of attention-based long short-term memory (LSTM) will prove to be efficient. Data Augmentation methods will be used for getting the better results. TensorFlow and Keras are the python libraries that will be used to convert the speech to text. Currently there are many software available but all requires the network connectivity for it to work, while this device will work without the requirement of internet.Using the proposed model we got the accuracy of 100% in predicting sign language and 96% accuracy in sentence level understanding.","PeriodicalId":135318,"journal":{"name":"2021 International Conference on Artificial Intelligence and Machine Vision (AIMV)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Artificial Intelligence and Machine Vision (AIMV)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/aimv53313.2021.9670928","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Deaf people may get irritated due to the problem of not being able to share their views with common people, which may affect their day-to-day life. This is the main reason to develop such system that can help these people and they can also put their thoughts forward similar to other people who don’t have such problem. The advancement in the Artificial intelligence provides the door for developing the system that overcome this difficulty. So this project aims on developing a system which will be able to convert the speech to text for the deaf person, and also sometimes the person might not be able to understand just by text, so the speech will also get converted to the universal sign language. Similarly, for the mute people the sign language which they are using will get converted to speech. We will take help of various ML and AI concepts along with NLP to develop the accurate model. Convolutional neural networks (CNN) will be used for prediction as it is efficient in predicting image input, also as lip movements are fast and continuous so it is hard to capture so along with CNN, the use of attention-based long short-term memory (LSTM) will prove to be efficient. Data Augmentation methods will be used for getting the better results. TensorFlow and Keras are the python libraries that will be used to convert the speech to text. Currently there are many software available but all requires the network connectivity for it to work, while this device will work without the requirement of internet.Using the proposed model we got the accuracy of 100% in predicting sign language and 96% accuracy in sentence level understanding.
实时手语转换为哑巴和聋人
聋人可能会因为无法与普通人分享他们的观点而感到愤怒,这可能会影响他们的日常生活。这就是开发这样一个系统的主要原因,它可以帮助这些人,他们也可以把他们的想法提出,就像其他没有这种问题的人一样。人工智能的进步为开发克服这一困难的系统提供了大门。所以这个项目的目标是开发一个系统,能够将语音转换为文本给聋哑人,有时候聋哑人可能无法通过文本理解,所以语音也会被转换为通用的手语。同样,对于哑巴来说,他们使用的手语也会被转化为语言。我们将利用各种ML和AI概念以及NLP来开发准确的模型。卷积神经网络(CNN)将被用于预测,因为它在预测图像输入方面是有效的,也因为嘴唇运动是快速和连续的,所以很难捕捉到,所以与CNN一起,使用基于注意力的长短期记忆(LSTM)将被证明是有效的。为了获得更好的结果,将使用数据增强方法。TensorFlow和Keras是将用于将语音转换为文本的python库。目前有很多软件可用,但都需要网络连接才能工作,而这个设备不需要互联网就可以工作。使用该模型,手语预测准确率达到100%,句子理解准确率达到96%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信