S. Akhter, Shah Jafor Sadeek Quaderi, Saleh Ud-Din Ahmed
{"title":"Deep Learning with OBH for Real-Time Rotation-Invariant Signs Detection","authors":"S. Akhter, Shah Jafor Sadeek Quaderi, Saleh Ud-Din Ahmed","doi":"10.1145/3587828.3587884","DOIUrl":null,"url":null,"abstract":"Numerous studies are being undertaken to provide answers for sign language recognition and classification. Deep learning-based models have higher accuracy (90%-98%); however, require more runtime memory and processing in terms of both computational power and execution time (1 hour 20 minutes) for feature extraction and training images. Besides, deep learning models are not entirely insensitive to translation, rotation, and scaling; unless the training data includes rotated, translated, or scaled signs. However, Orientation-Based Hashcode (OBH) completes gesture recognition in a significantly shorter length of time (5 minutes) and with reasonable accuracy (80%-85%). In addition, OBH is not affected by translation, rotation, scaling, or occlusion. As a result, a new intermediary model is developed to detect sign language and perform classification with a reasonable processing time (6 minutes) like OBH while providing attractive accuracy (90%-96%) and invariance qualities. This paper presents a coupled and completely networked autonomous system comprised of OBH and Gabor features with machine learning models. The proposed model is evaluated with 576 sign alphabet images (RGB and Depth) from 24 distinct categories, and the results are compared to those obtained using traditional machine learning methodologies. The proposed methodology is 95.8% accurate against a randomly selected test dataset and 93.85% accurate after 9-fold validation.","PeriodicalId":340917,"journal":{"name":"Proceedings of the 2023 12th International Conference on Software and Computer Applications","volume":"192 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2023 12th International Conference on Software and Computer Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3587828.3587884","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Numerous studies are being undertaken to provide answers for sign language recognition and classification. Deep learning-based models have higher accuracy (90%-98%); however, require more runtime memory and processing in terms of both computational power and execution time (1 hour 20 minutes) for feature extraction and training images. Besides, deep learning models are not entirely insensitive to translation, rotation, and scaling; unless the training data includes rotated, translated, or scaled signs. However, Orientation-Based Hashcode (OBH) completes gesture recognition in a significantly shorter length of time (5 minutes) and with reasonable accuracy (80%-85%). In addition, OBH is not affected by translation, rotation, scaling, or occlusion. As a result, a new intermediary model is developed to detect sign language and perform classification with a reasonable processing time (6 minutes) like OBH while providing attractive accuracy (90%-96%) and invariance qualities. This paper presents a coupled and completely networked autonomous system comprised of OBH and Gabor features with machine learning models. The proposed model is evaluated with 576 sign alphabet images (RGB and Depth) from 24 distinct categories, and the results are compared to those obtained using traditional machine learning methodologies. The proposed methodology is 95.8% accurate against a randomly selected test dataset and 93.85% accurate after 9-fold validation.