K.Bhanu Prathap, G.Divya Swaroop, B.Praveen Kumar, V. Kamble, Mayuri A Parate
{"title":"印度手语识别","authors":"K.Bhanu Prathap, G.Divya Swaroop, B.Praveen Kumar, V. Kamble, Mayuri A Parate","doi":"10.1109/PCEMS58491.2023.10136062","DOIUrl":null,"url":null,"abstract":"Normal people can readily connect and communicate with one another, however, those with hearing and speech impairments find it difficult to converse with normal-hearing people without the assistance of a translator. The only way deaf and dumb people can communicate is through Sign Language. Indian Sign Language has its own grammar, syntax, vocabulary, and unique language features. We propose two methods, namely Bidirectional LSTM and BERT Transformer to address the problem of sign language translation. The proposed work is validated on standard datasets and provides promising results. The INCLUDE-50 dataset is used to validate the performance of the proposed algorithm. The deep neural network is evaluated using a combination of approaches for augmentation of the data, features extraction using the mediapipe.On the Dataset INCLUDE 50 the best performing model obtained an accuracy of 89.5%. This model employs a feature extractor that has been pre-trained, as well as an encoder and a decoder.","PeriodicalId":330870,"journal":{"name":"2023 2nd International Conference on Paradigm Shifts in Communications Embedded Systems, Machine Learning and Signal Processing (PCEMS)","volume":"568 ","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"ISLR: Indian Sign Language Recognition\",\"authors\":\"K.Bhanu Prathap, G.Divya Swaroop, B.Praveen Kumar, V. Kamble, Mayuri A Parate\",\"doi\":\"10.1109/PCEMS58491.2023.10136062\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Normal people can readily connect and communicate with one another, however, those with hearing and speech impairments find it difficult to converse with normal-hearing people without the assistance of a translator. The only way deaf and dumb people can communicate is through Sign Language. Indian Sign Language has its own grammar, syntax, vocabulary, and unique language features. We propose two methods, namely Bidirectional LSTM and BERT Transformer to address the problem of sign language translation. The proposed work is validated on standard datasets and provides promising results. The INCLUDE-50 dataset is used to validate the performance of the proposed algorithm. The deep neural network is evaluated using a combination of approaches for augmentation of the data, features extraction using the mediapipe.On the Dataset INCLUDE 50 the best performing model obtained an accuracy of 89.5%. This model employs a feature extractor that has been pre-trained, as well as an encoder and a decoder.\",\"PeriodicalId\":330870,\"journal\":{\"name\":\"2023 2nd International Conference on Paradigm Shifts in Communications Embedded Systems, Machine Learning and Signal Processing (PCEMS)\",\"volume\":\"568 \",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-04-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 2nd International Conference on Paradigm Shifts in Communications Embedded Systems, Machine Learning and Signal Processing (PCEMS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/PCEMS58491.2023.10136062\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 2nd International Conference on Paradigm Shifts in Communications Embedded Systems, Machine Learning and Signal Processing (PCEMS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/PCEMS58491.2023.10136062","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
正常人可以很容易地相互联系和交流,然而,那些有听力和语言障碍的人发现,如果没有翻译的帮助,很难与听力正常的人交谈。聋哑人交流的唯一方式是通过手语。印度手语有自己的语法、句法、词汇和独特的语言特征。我们提出了两种方法,即双向LSTM和BERT转换器来解决手语翻译问题。所提出的工作在标准数据集上进行了验证,并提供了有希望的结果。使用INCLUDE-50数据集验证了所提出算法的性能。深度神经网络的评估使用了数据增强和mediapipe特征提取的组合方法。在Dataset INCLUDE 50上,表现最好的模型获得了89.5%的准确率。该模型采用了一个预训练的特征提取器,以及一个编码器和一个解码器。
Normal people can readily connect and communicate with one another, however, those with hearing and speech impairments find it difficult to converse with normal-hearing people without the assistance of a translator. The only way deaf and dumb people can communicate is through Sign Language. Indian Sign Language has its own grammar, syntax, vocabulary, and unique language features. We propose two methods, namely Bidirectional LSTM and BERT Transformer to address the problem of sign language translation. The proposed work is validated on standard datasets and provides promising results. The INCLUDE-50 dataset is used to validate the performance of the proposed algorithm. The deep neural network is evaluated using a combination of approaches for augmentation of the data, features extraction using the mediapipe.On the Dataset INCLUDE 50 the best performing model obtained an accuracy of 89.5%. This model employs a feature extractor that has been pre-trained, as well as an encoder and a decoder.