Interpretation of Bahasa Isyarat Malaysia (BIM) Using SSD-MobileNet-V2 FPNLite and COCO mAP

Inf. Comput. Pub Date : 2023-05-31 DOI:10.3390/info14060319
Iffah Zulaikha Saiful Bahri, S. Saon, A. Mahamad, K. Isa, U. Fadlilah, Mohd Anuaruddin Bin Ahmadon, S. Yamaguchi
{"title":"Interpretation of Bahasa Isyarat Malaysia (BIM) Using SSD-MobileNet-V2 FPNLite and COCO mAP","authors":"Iffah Zulaikha Saiful Bahri, S. Saon, A. Mahamad, K. Isa, U. Fadlilah, Mohd Anuaruddin Bin Ahmadon, S. Yamaguchi","doi":"10.3390/info14060319","DOIUrl":null,"url":null,"abstract":"This research proposes a study on two-way communication between deaf/mute and normal people using an Android application. Despite advancements in technology, there is still a lack of mobile applications that facilitate two-way communication between deaf/mute and normal people, especially by using Bahasa Isyarat Malaysia (BIM). This project consists of three parts: First, we use BIM letters, which enables the recognition of BIM letters and BIM combined letters to form a word. In this part, a MobileNet pre-trained model is implemented to train the model with a total of 87,000 images for 29 classes, with a 10% test size and a 90% training size. The second part is BIM word hand gestures, which consists of five classes that are trained with the SSD-MobileNet-V2 FPNLite 320 × 320 pre-trained model with a speed of 22 s/frame rate and COCO mAP of 22.2, with a total of 500 images for all five classes and first-time training set to 2000 steps, while the second- and third-time training are set to 2500 steps. The third part is Android application development using Android Studio, which contains the features of the BIM letters and BIM word hand gestures, with the trained models converted into TensorFlow Lite. This feature also includes the conversion of speech to text, whereby this feature allows converting speech to text through the Android application. Thus, BIM letters obtain 99.75% accuracy after training the models, while BIM word hand gestures obtain 61.60% accuracy. The suggested system is validated as a result of these simulations and tests.","PeriodicalId":13622,"journal":{"name":"Inf. Comput.","volume":"5 1","pages":"319"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Inf. Comput.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3390/info14060319","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

This research proposes a study on two-way communication between deaf/mute and normal people using an Android application. Despite advancements in technology, there is still a lack of mobile applications that facilitate two-way communication between deaf/mute and normal people, especially by using Bahasa Isyarat Malaysia (BIM). This project consists of three parts: First, we use BIM letters, which enables the recognition of BIM letters and BIM combined letters to form a word. In this part, a MobileNet pre-trained model is implemented to train the model with a total of 87,000 images for 29 classes, with a 10% test size and a 90% training size. The second part is BIM word hand gestures, which consists of five classes that are trained with the SSD-MobileNet-V2 FPNLite 320 × 320 pre-trained model with a speed of 22 s/frame rate and COCO mAP of 22.2, with a total of 500 images for all five classes and first-time training set to 2000 steps, while the second- and third-time training are set to 2500 steps. The third part is Android application development using Android Studio, which contains the features of the BIM letters and BIM word hand gestures, with the trained models converted into TensorFlow Lite. This feature also includes the conversion of speech to text, whereby this feature allows converting speech to text through the Android application. Thus, BIM letters obtain 99.75% accuracy after training the models, while BIM word hand gestures obtain 61.60% accuracy. The suggested system is validated as a result of these simulations and tests.
使用SSD-MobileNet-V2 FPNLite和COCO mAP的马来文BIM解释
本研究提出使用Android应用程序对聋哑人与正常人之间的双向交流进行研究。尽管技术进步,但仍然缺乏促进聋哑人与正常人之间双向交流的移动应用程序,特别是使用马来西亚语(BIM)。本项目由三部分组成:首先,我们使用BIM字母,可以识别BIM字母和BIM组合字母组成一个单词。在这一部分中,我们实现了一个MobileNet预训练模型,对29个类共87000张图像的模型进行训练,测试大小为10%,训练大小为90%。第二部分是BIM单词手势,由五个类组成,使用SSD-MobileNet-V2 FPNLite 320 × 320预训练模型进行训练,速度为22 s/帧率,COCO mAP为22.2,五个类总共500张图像,第一次训练设置为2000步,第二次和第三次训练设置为2500步。第三部分是使用Android Studio进行Android应用程序开发,其中包含BIM字母和BIM单词手势的特征,并将训练好的模型转换为TensorFlow Lite。该功能还包括语音到文本的转换,即该功能允许通过Android应用程序将语音转换为文本。因此,经过模型训练后,BIM字母的准确率达到99.75%,而BIM单词手势的准确率达到61.60%。通过仿真和测试,验证了该系统的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信