Jetson Nano-Based Two-Way Communication System with Filipino Sign Language Recognition Using LSTM Deep Learning Model for Able and Deaf-Mute Persons

Rain Kristine B. Cabigting, Carl James U. Grantoza, Leonardo D. Valiente, Ericson D. Dimaunahan
{"title":"Jetson Nano-Based Two-Way Communication System with Filipino Sign Language Recognition Using LSTM Deep Learning Model for Able and Deaf-Mute Persons","authors":"Rain Kristine B. Cabigting, Carl James U. Grantoza, Leonardo D. Valiente, Ericson D. Dimaunahan","doi":"10.1109/RAAI56146.2022.10092971","DOIUrl":null,"url":null,"abstract":"Communication is the foundation of what it is to be human. The majority of human communication is reliant on sounds. However, it is not the sole natural means of communication; other people employ alternative ways. One of which is the Deaf community’s language. Communication between the Deaf-mute community and hearing or able individuals is one of the various challenges the two parties encounter. In the Philippines, around 70% of the Filipino Deaf Community utilizes Filipino Sign Language (FSL) as their primary language, whereas some hearing persons may be illiterate in the Deaf community’s native language. With the given situation, this study created a twoway communication device using Jetson Nano, covering the proper translation of FSL to text and speech and the conversion of input speech into text. Ten (10) dynamic FSL gestures are considered in this study. The device used LSTM Deep Learning Model and MediaPipe to recognize the FSL gestures, then convert them into speech through Google Text-to-Speech (gTTS) API. The device also converts speech to text using Google Speech-to-Text (gSTT) API. Sixty (60) trials of a two-way conversation between a deaf-mute and a hearing person are performed. By conducting a Test of Proportion for a Two-Way Conversation, it has revealed that the prototype exceeded the standard value of 91.11%, which garnered an accuracy of 93.33%, rendering the device highly effective and reliable as a means of communication between a deaf-mute person and a fully able one.","PeriodicalId":190255,"journal":{"name":"2022 2nd International Conference on Robotics, Automation and Artificial Intelligence (RAAI)","volume":"69 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 2nd International Conference on Robotics, Automation and Artificial Intelligence (RAAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/RAAI56146.2022.10092971","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Communication is the foundation of what it is to be human. The majority of human communication is reliant on sounds. However, it is not the sole natural means of communication; other people employ alternative ways. One of which is the Deaf community’s language. Communication between the Deaf-mute community and hearing or able individuals is one of the various challenges the two parties encounter. In the Philippines, around 70% of the Filipino Deaf Community utilizes Filipino Sign Language (FSL) as their primary language, whereas some hearing persons may be illiterate in the Deaf community’s native language. With the given situation, this study created a twoway communication device using Jetson Nano, covering the proper translation of FSL to text and speech and the conversion of input speech into text. Ten (10) dynamic FSL gestures are considered in this study. The device used LSTM Deep Learning Model and MediaPipe to recognize the FSL gestures, then convert them into speech through Google Text-to-Speech (gTTS) API. The device also converts speech to text using Google Speech-to-Text (gSTT) API. Sixty (60) trials of a two-way conversation between a deaf-mute and a hearing person are performed. By conducting a Test of Proportion for a Two-Way Conversation, it has revealed that the prototype exceeded the standard value of 91.11%, which garnered an accuracy of 93.33%, rendering the device highly effective and reliable as a means of communication between a deaf-mute person and a fully able one.
基于Jetson纳米的菲律宾手语双向交流系统,使用LSTM深度学习模型,用于残疾人和聋哑人
沟通是做人的基础。人类的大部分交流都依赖于声音。然而,它并不是唯一的自然交流方式;其他人则采用其他方法。其中之一是聋人社区的语言。聋哑人社区与听力健全者之间的沟通是双方面临的各种挑战之一。在菲律宾,大约70%的菲律宾聋人社区使用菲律宾手语(FSL)作为他们的主要语言,而一些听力正常的人可能不懂聋人社区的母语。在给定的情况下,本研究使用Jetson Nano创建了一个双向通信设备,包括FSL到文本和语音的正确翻译以及输入语音到文本的转换。本研究考虑了十(10)种动态FSL手势。该设备使用LSTM深度学习模型和MediaPipe识别FSL手势,然后通过Google文本到语音(gTTS) API将其转换为语音。该设备还使用谷歌语音到文本(gSTT) API将语音转换为文本。在聋哑人和听力正常的人之间进行了60次双向对话试验。通过双向对话比例测试,样机超过了91.11%的标准值,准确率达到了93.33%,作为聋哑人与正常人之间的交流工具,该设备是非常有效和可靠的。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信