{"title":"Leveraging edge detection techniques to enhance Arabic sign language static-gesture recognition using deep learning","authors":"Wahiba Ismaiel , Lilia kechiche , Yassine Aribi , Omer Salih Dawood Omer , Walied Merghani","doi":"10.1016/j.jer.2025.09.011","DOIUrl":null,"url":null,"abstract":"<div><div>Recently, there has been a growing interest in developing solutions to address communication barriers for the deaf and hard-of-hearing community. Sign language is the primary language of this group. Computer vision technology is used to process sign language due to its ease of application. Sign language recognition involves the use of technology to bridge communication gaps and enhance accessibility for individuals who use sign language as their primary form of communication. Many researchers have presented various methods to facilitate communication, among others. These methods include sign language recognition techniques, translation between sign and text or audio, and hand gesture identification, among others. We proposed an effective approach to improve the feature extraction process for Arabic sign gesture recognition. Feature extraction is a crucial aspect of deep learning models because it facilitates data processing, improves performance, and helps interpret results. This process also enables models to manage large datasets more efficiently. We presented two deep learning models: the agile convolutional neural network (ASLR_CNN) and ResNet50, to improve the comprehensiveness of the extracted features. These models were combined with the Canny Edge Detector (CED), which identifies the edges of Arabic hand gestures, as well as the complex features extracted from the edges by the proposed models. To evaluate the effectiveness of our methodology, we trained the proposed models on two public datasets: AASL and ArASL. The performance of these models was evaluated using a variety of metrics, including accuracy, precision, recall, F-score, and confusion matrix. The results indicated that both the ASLR_CNN and ResNet50 models achieved high accuracy on the ArASL dataset, reaching 97.14 % and 96.88 %, respectively. In contrast, the accuracy dropped to 89.49 % and 86.12 % for the ASLR_CNN and ResNet50 models, respectively, when using the AASL dataset.</div></div>","PeriodicalId":48803,"journal":{"name":"Journal of Engineering Research","volume":"14 1","pages":"Pages 896-915"},"PeriodicalIF":2.2000,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Engineering Research","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2307187725005711","RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/9/24 0:00:00","PubModel":"Epub","JCR":"Q3","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0
Abstract
Recently, there has been a growing interest in developing solutions to address communication barriers for the deaf and hard-of-hearing community. Sign language is the primary language of this group. Computer vision technology is used to process sign language due to its ease of application. Sign language recognition involves the use of technology to bridge communication gaps and enhance accessibility for individuals who use sign language as their primary form of communication. Many researchers have presented various methods to facilitate communication, among others. These methods include sign language recognition techniques, translation between sign and text or audio, and hand gesture identification, among others. We proposed an effective approach to improve the feature extraction process for Arabic sign gesture recognition. Feature extraction is a crucial aspect of deep learning models because it facilitates data processing, improves performance, and helps interpret results. This process also enables models to manage large datasets more efficiently. We presented two deep learning models: the agile convolutional neural network (ASLR_CNN) and ResNet50, to improve the comprehensiveness of the extracted features. These models were combined with the Canny Edge Detector (CED), which identifies the edges of Arabic hand gestures, as well as the complex features extracted from the edges by the proposed models. To evaluate the effectiveness of our methodology, we trained the proposed models on two public datasets: AASL and ArASL. The performance of these models was evaluated using a variety of metrics, including accuracy, precision, recall, F-score, and confusion matrix. The results indicated that both the ASLR_CNN and ResNet50 models achieved high accuracy on the ArASL dataset, reaching 97.14 % and 96.88 %, respectively. In contrast, the accuracy dropped to 89.49 % and 86.12 % for the ASLR_CNN and ResNet50 models, respectively, when using the AASL dataset.
期刊介绍:
Journal of Engineering Research (JER) is a international, peer reviewed journal which publishes full length original research papers, reviews, case studies related to all areas of Engineering such as: Civil, Mechanical, Industrial, Electrical, Computer, Chemical, Petroleum, Aerospace, Architectural, Biomedical, Coastal, Environmental, Marine & Ocean, Metallurgical & Materials, software, Surveying, Systems and Manufacturing Engineering. In particular, JER focuses on innovative approaches and methods that contribute to solving the environmental and manufacturing problems, which exist primarily in the Arabian Gulf region and the Middle East countries. Kuwait University used to publish the Journal "Kuwait Journal of Science and Engineering" (ISSN: 1024-8684), which included Science and Engineering articles since 1974. In 2011 the decision was taken to split KJSE into two independent Journals - "Journal of Engineering Research "(JER) and "Kuwait Journal of Science" (KJS).