基于EfficientnetB1和迁移学习技术的阿拉伯手语识别

Basel A. Dabwan, M. Jadhav, Yahya A. Ali, Fekry Olayah
{"title":"基于EfficientnetB1和迁移学习技术的阿拉伯手语识别","authors":"Basel A. Dabwan, M. Jadhav, Yahya A. Ali, Fekry Olayah","doi":"10.1109/ITIKD56332.2023.10099710","DOIUrl":null,"url":null,"abstract":"Deaf and mute people rely on signing as a means of communication with others and with themselves. The value of sign language, which is the sole way for the deaf and mute communities to communicate, is often overlooked by regular people. Because of these limitations or impairments, these people are experiencing considerable setbacks in their life, including joblessness, serious depression, and a variety of other symptoms. Sign language interpreters are one of the communication services they use. However, paying these interpreters is prohibitively expensive, necessitating a low-cost solution to the problem. As a result, a system has been built that would interpret the Arabic Sign Language-based visual hand dataset into written information. The dataset used for this model Arabic Alphabets Sign Language dataset consists of 32 classes, each category has 506 images, resulting in a total of 506 * 32 = 16192 images. On the provided dataset, the tests have been run using a variety of pre-trained models. Most of them carried out their duties rather normally, and in the end, we constructed the Convolutional Neural Networks model with EfficientnetB1 scaling loaded with weights pre-trained on the ImageNet model, Based on this observation. Using a simple but highly effective compound coefficient, it equally scales all width/depth/resolution dimensions. Based on the model's results, we demonstrated the efficacy of this strategy. The percentage of accuracy that we obtained from this model is 99% accuracy and 97.9% validation accuracy","PeriodicalId":283631,"journal":{"name":"2023 International Conference on IT Innovation and Knowledge Discovery (ITIKD)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Arabic Sign Language Recognition Using EfficientnetB1 and Transfer Learning Technique\",\"authors\":\"Basel A. Dabwan, M. Jadhav, Yahya A. Ali, Fekry Olayah\",\"doi\":\"10.1109/ITIKD56332.2023.10099710\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Deaf and mute people rely on signing as a means of communication with others and with themselves. The value of sign language, which is the sole way for the deaf and mute communities to communicate, is often overlooked by regular people. Because of these limitations or impairments, these people are experiencing considerable setbacks in their life, including joblessness, serious depression, and a variety of other symptoms. Sign language interpreters are one of the communication services they use. However, paying these interpreters is prohibitively expensive, necessitating a low-cost solution to the problem. As a result, a system has been built that would interpret the Arabic Sign Language-based visual hand dataset into written information. The dataset used for this model Arabic Alphabets Sign Language dataset consists of 32 classes, each category has 506 images, resulting in a total of 506 * 32 = 16192 images. On the provided dataset, the tests have been run using a variety of pre-trained models. Most of them carried out their duties rather normally, and in the end, we constructed the Convolutional Neural Networks model with EfficientnetB1 scaling loaded with weights pre-trained on the ImageNet model, Based on this observation. Using a simple but highly effective compound coefficient, it equally scales all width/depth/resolution dimensions. Based on the model's results, we demonstrated the efficacy of this strategy. The percentage of accuracy that we obtained from this model is 99% accuracy and 97.9% validation accuracy\",\"PeriodicalId\":283631,\"journal\":{\"name\":\"2023 International Conference on IT Innovation and Knowledge Discovery (ITIKD)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-03-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 International Conference on IT Innovation and Knowledge Discovery (ITIKD)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ITIKD56332.2023.10099710\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 International Conference on IT Innovation and Knowledge Discovery (ITIKD)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ITIKD56332.2023.10099710","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

聋哑人依靠手语作为与他人和自己交流的一种手段。手语是聋哑人交流的唯一方式,它的价值常常被普通人所忽视。由于这些限制或损伤,这些人在生活中经历了相当大的挫折,包括失业、严重抑郁和各种其他症状。手语翻译是他们使用的交流服务之一。然而,支付这些口译员的费用非常昂贵,因此需要一个低成本的解决方案来解决这个问题。因此,已经建立了一个系统,可以将基于阿拉伯手语的视觉手部数据集解释为书面信息。本模型使用的数据集由32个类组成,每个类有506张图片,总共有506 * 32 = 16192张图片。在提供的数据集上,使用各种预训练的模型运行了测试。他们中的大多数人都很正常地履行了自己的职责,最后,我们基于这一观察,构建了基于ImageNet模型预训练权重的卷积神经网络模型。使用简单但高效的复合系数,它可以平等地缩放所有宽度/深度/分辨率维度。基于模型的结果,我们证明了该策略的有效性。我们从该模型中获得的准确率为99%,验证准确率为97.9%
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Arabic Sign Language Recognition Using EfficientnetB1 and Transfer Learning Technique
Deaf and mute people rely on signing as a means of communication with others and with themselves. The value of sign language, which is the sole way for the deaf and mute communities to communicate, is often overlooked by regular people. Because of these limitations or impairments, these people are experiencing considerable setbacks in their life, including joblessness, serious depression, and a variety of other symptoms. Sign language interpreters are one of the communication services they use. However, paying these interpreters is prohibitively expensive, necessitating a low-cost solution to the problem. As a result, a system has been built that would interpret the Arabic Sign Language-based visual hand dataset into written information. The dataset used for this model Arabic Alphabets Sign Language dataset consists of 32 classes, each category has 506 images, resulting in a total of 506 * 32 = 16192 images. On the provided dataset, the tests have been run using a variety of pre-trained models. Most of them carried out their duties rather normally, and in the end, we constructed the Convolutional Neural Networks model with EfficientnetB1 scaling loaded with weights pre-trained on the ImageNet model, Based on this observation. Using a simple but highly effective compound coefficient, it equally scales all width/depth/resolution dimensions. Based on the model's results, we demonstrated the efficacy of this strategy. The percentage of accuracy that we obtained from this model is 99% accuracy and 97.9% validation accuracy
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信